⚡️ More Reasonable LLMs

PLUS: ChatGPT tricked into revealing personal data

Good Morning. Meta breathes new life into AI reasoning with its innovative System 2 Attention. But will this approach herald a new era for accurate AI, or will increased complexity prove a stumbling block? Let’s dive in.

Today’s Highlights:

  • Stability AI's turbulent times and technical triumphs

  • ChatGPT tricked into revealing personal data

  • DeepMind’s GNoME AI identifies millions of crystals

DEEP DIVE

Meta's System 2 Attention Boosts LLM Reasoning Ability

Meta’s new breakthrough presents a significant advance for language models. Dubbed System 2 Attention (S2A), the technique enhances the problem-solving acumen of LLMs by filtering out noise and honing in on the key information needed for sound reasoning.

LLMs often trip over extraneous bits in prompts, like personal guesses or unrelated facts. This can lean their responses towards confirmation instead of accuracy.

LLMs use contextual information to produce output, even if it is irrelevant

Traditionally, transformers—the neural networks underpinning LLMs—propel this issue forward by training on next-token prediction, which makes them highly context-sensitive. Unfortunately, this sensitivity sometimes means that irrelevant information gets undue attention, skewing the reasoning process.

What is System 2 Attention?

S2A introduces a refined attention mechanism by trimming the fat from the context before an LLM crafts its response. It's akin to Daniel Kahneman's System 2 thinking—slow, deliberate, logical.

Here’s how S2A works:

  1. First, S2A simplifies the context by stripping out irrelevant elements.

  2. Then, this streamlined context is handed over to the primary LLM to generate a response.

    S2A removes irrelevant information from the prompt to improve accuracy

This process doesn't just improve the quality of the output; it also guides LLMs away from biases and toward factual, objective answers, especially in question-answering and long-form generation.

The trade-off? While S2A results are promising, there's an acknowledged increase in computation and complexity. More steps in the generation process and the requirement to sift through the initial prompt mean more resources.

Despite these limitations, the technique signals a meaningful advancement, potentially bolstering the use of LLMs in situations where precise reasoning is non-negotiable.

PUNCHLINES

Diffusing the Situation: Stability AI weathers internal turmoil and seeks shelter under a potential acquisition as the CEO faces scrutiny.

Fresh Search on the Block: Ex-Google team's Perplexity AI launches LLMs to usurp the search throne.

The AI Underworld: Cybercriminals are now using ChatGPT to polish phishing scams to a deceptive sheen.

MS Paint Gets a New Artist: The classic Paint app gets an AI upgrade, inviting users to sketch out their thoughts with DALL-E's brainpower.

Pixels at Work: Amazon's Titan Image Generator leaps into the AI art arena with enterprise armor and legal shields.

TLDR

ChatGPT exposes sensitive data through simple prompt: Researchers tricked ChatGPT into revealing personal data like email addresses and phone numbers with a basic command to repeat words indefinitely. The incident raises concerns about the training data of AI models and prompts calls for stricter testing protocols.

Stable Diffusion XL Turbo enhances real-time AI art: Stability AI upgrades its text-to-image AI, introducing SDXL Turbo with Adversarial Diffusion Distillation, cutting down image generation from 50 steps to just one. SDXL Turbo promises GAN-like speed without quality loss, topping its predecessors in blind tests and producing 512x512 images in 207ms on Nvidia A100 GPUs.

DeepMind's GNoME AI reveals over 2 million new crystals: DeepMind's GNoME AI, in collaboration with Berkeley Lab, has identified 2.2 million new crystal structures, boosting the known stable inorganic crystals by almost 10x. This achievement, tested with a 71% success rate using autonomous labs, may significantly advance technologies like batteries and solar panels.

AWS introduces benchmarking for AI models: AWS launches Model Evaluation on Bedrock, combining automated and human evaluations for AI models. The service enhances selection precision with performance reports, recognizes issues like toxicity, and offers human evaluators for nuanced metrics such as empathy.

TRENDING TOOLS

💻 AiTerm: Transform natural language into commands with an AI-enhanced terminal assistant

🕸️ Usescarper: An AI-powered web crawler for efficient data extraction

🐒 JungleGym: An open-source framework for crafting self-governing web bots

🌐 Meshy: Create stunning 3D content from text and images in seconds

🤥 LiarLiar: Utilize AI to discern lies and monitor heart rate variability

That’s all for today—if you have any questions or something interesting to share, please reply to this email. We’d love to hear from you!

P.S. If you want to sign up for the Supercharged newsletter or share it with a friend, you can find us here.

Reply

or to participate.