Close Menu
Must Have Gadgets –

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    More YouTube age verification checks are hitting viewers

    November 7, 2025

    Can’t Here a Thing? Here’s What to Do if Your Laptop Has No Sound

    November 7, 2025

    Here’s how much Apple will reportedly pay Google for Gemini to save Siri

    November 7, 2025
    Facebook X (Twitter) Instagram
    Must Have Gadgets –
    Trending
    • More YouTube age verification checks are hitting viewers
    • Can’t Here a Thing? Here’s What to Do if Your Laptop Has No Sound
    • Here’s how much Apple will reportedly pay Google for Gemini to save Siri
    • The Best Windows Mini PCs We’ve Tested for 2025
    • Pluribus review: Apple TV’s latest sci-fi series luxuriates in mystery
    • Today’s NYT Mini Crossword Answers for Nov. 7
    • Elon Musk says Tesla owners can ‘text and drive’ very soon
    • Pebblebee starts sales of Clip 5 and Card 5 Find Hub trackers
    • Home
    • Shop
      • Earbuds & Headphones
      • Smartwatches
      • Mobile Accessories
      • Smart Home Devices
      • Laptops & Tablets
    • Gadget Reviews
    • How-To Guides
    • Mobile Accessories
    • Smart Devices
    • More
      • Top Deals
      • Smart Home
      • Tech News
      • Trending Tech
    Facebook X (Twitter) Instagram
    Must Have Gadgets –
    Home»Tech News»LLMs show a “highly unreliable” capacity to describe their own internal processes
    Tech News

    LLMs show a “highly unreliable” capacity to describe their own internal processes

    adminBy adminNovember 4, 2025No Comments3 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    LLMs show a “highly unreliable” capacity to describe their own internal processes
    Share
    Facebook Twitter LinkedIn Pinterest Email

    WHY ARE WE ALL YELLING?!


    Credit:

    Anthropic

    Unfortunately for AI self-awareness boosters, this demonstrated ability was extremely inconsistent and brittle across repeated tests. The best-performing models in Anthropic’s tests—Opus 4 and 4.1—topped out at correctly identifying the injected concept just 20 percent of the time.

    In a similar test where the model was asked “Are you experiencing anything unusual?” Opus 4.1 improved to a 42 percent success rate that nonetheless still fell below even a bare majority of trials. The size of the “introspection” effect was also highly sensitive to which internal model layer the insertion was performed on—if the concept was introduced too early or too late in the multi-step inference process, the “self-awareness” effect disappeared completely.

    Show us the mechanism

    Anthropic also took a few other tacks to try to get an LLM’s understanding of its internal state. When asked to “tell me what word you’re thinking about” while reading an unrelated line, for instance, the models would sometimes mention a concept that had been injected into its activations. And when asked to defend a forced response matching an injected concept, the LLM would sometimes apologize and “confabulate an explanation for why the injected concept came to mind.” In every case, though, the result was highly inconsistent across multiple trials.

    Even the most “introspective” models tested by Anthropic only detected the injected “thoughts” about 20 percent of the time.

    Even the most “introspective” models tested by Anthropic only detected the injected “thoughts” about 20 percent of the time.


    Credit:

    Antrhopic

    In the paper, the researchers put some positive spin on the apparent fact that “current language models possess some functional introspective awareness of their own internal states” [emphasis added]. At the same time, they acknowledge multiple times that this demonstrated ability is much too brittle and context-dependent to be considered dependable. Still, Anthropic hopes that such features “may continue to develop with further improvements to model capabilities.”

    One thing that might stop such advancement, though, is an overall lack of understanding of the precise mechanism leading to these demonstrated “self-awareness” effects. The researchers theorize about “anomaly detection mechanisms” and “consistency-checking circuits” that might develop organically during the training process to “effectively compute a function of its internal representations” but don’t settle on any concrete explanation.

    In the end, it will take further research to understand how, exactly, an LLM even begins to show any understanding about how it operates. For now, the researchers acknowledge, “the mechanisms underlying our results could still be rather shallow and narrowly specialized.” And even then, they hasten to add that these LLM capabilities “may not have the same philosophical significance they do in humans, particularly given our uncertainty about their mechanistic basis.”

    capacity describe highly Internal LLMs processes show unreliable
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    admin
    • Website

    Related Posts

    Elon Musk says Tesla owners can ‘text and drive’ very soon

    November 7, 2025

    Elon Musk wins $1 trillion Tesla pay vote despite “part-time CEO” criticism

    November 7, 2025

    Tesla delays reveal of production Roadster 2 to April Fools’ Day

    November 7, 2025
    Leave A Reply Cancel Reply

    Top Posts

    More YouTube age verification checks are hitting viewers

    November 7, 2025

    PayPal’s blockchain partner accidentally minted $300 trillion in stablecoins

    October 16, 2025

    The best AirPods deals for October 2025

    October 16, 2025
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews
    How-To Guides

    How to Disable Some or All AI Features on your Samsung Galaxy Phone

    By adminOctober 16, 20250
    Gadget Reviews

    PayPal’s blockchain partner accidentally minted $300 trillion in stablecoins

    By adminOctober 16, 20250
    Smart Devices

    The best AirPods deals for October 2025

    By adminOctober 16, 20250

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Latest Post

    More YouTube age verification checks are hitting viewers

    November 7, 2025

    Can’t Here a Thing? Here’s What to Do if Your Laptop Has No Sound

    November 7, 2025

    Here’s how much Apple will reportedly pay Google for Gemini to save Siri

    November 7, 2025
    Recent Posts
    • More YouTube age verification checks are hitting viewers
    • Can’t Here a Thing? Here’s What to Do if Your Laptop Has No Sound
    • Here’s how much Apple will reportedly pay Google for Gemini to save Siri
    • The Best Windows Mini PCs We’ve Tested for 2025
    • Pluribus review: Apple TV’s latest sci-fi series luxuriates in mystery

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms and Conditions
    • Disclaimer
    © 2025 must-have-gadgets.

    Type above and press Enter to search. Press Esc to cancel.