Close Menu
Pineapples Update –Pineapples Update –

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    I tried 0patch as a last resort for my Windows 10 PC – here’s how it compares to its promises

    January 20, 2026

    A PC Expert Explains Why Don’t Use Your Router’s USB Port When These Options Are Present

    January 20, 2026

    New ‘Remote Labor Index’ shows AI fails 97% of the time in freelancer tasks

    January 19, 2026
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Pineapples Update –Pineapples Update –
    • Home
    • Gaming
    • Gadgets
    • Startups
    • Security
    • How-To
    • AI/ML
    • Apps
    • Web3
    Pineapples Update –Pineapples Update –
    Home»How-To»Chatgpt is becoming smarter, but its hallucinations are spilling
    How-To

    Chatgpt is becoming smarter, but its hallucinations are spilling

    PineapplesUpdateBy PineapplesUpdateMay 7, 2025No Comments4 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Chatgpt is becoming smarter, but its hallucinations are spilling
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Chatgpt is becoming smarter, but its hallucinations are spilling


    • The latest AI models of Openai, GPT O3 and O4-Min, more often than their predecessors
    • Increased complexity of models may lead to more confidence impurities.
    • High error rate raises concerns about AI reliability in real -world applications

    Fantastic but incredible people are a major part of the story (and history). Based on an inquiry by Openai and, the same correlation may apply to AI Shared By the new York TimesHappiness, fictional facts and straightforward lies have been part of the AI ​​Chatbot as they were created. Theoretically improvement in model should reduce the frequency with which they appear.

    Openai’s latest flagship models, GPT O3 and O4-Min are to mimic human logic. Unlike its predecessors, who mainly focused on the fluent text generation, Openai built the GPT O3 and O 4-Mune to think of things through step-by-step. Openai has claimed that O1 may match or more than the performance of PhD students in Chemistry, Biology and Mathematics. But Openai’s report highlights some harsh results for any person that takes the reactions of chat at the inscribed price.

    Openai found that the GPT O3 model included hallucinations in one third of benchmark testing including public data. This is double the error rate of the O1 model before the previous year. The more compact O4-Mini model performed worse, hallucinations at 48% of similar tasks.

    You may like

    When more general knowledge questions are tested for the simplqa benchmark, the hallucinations give 51% reactions to the O3 and 79% reaction to the O4-Min. This is not just a little noise in the system; This is a fully developed identity crisis. You feel that the marketing as an argument system would re -check your own argument before creating some at least one answer, but this is not just the case.

    One of the principles making round in the AI ​​research community is that the more a model tries to argue, the more likely it is to move away from the rail. Unlike simple models, which stick to high-confidence predictions, logic models enter the field, where they should evaluate many possible routes, connect unequal facts, and essentially improve. And correction around the facts is also known as making things.

    Imaginary work

    Correlation is not the reason, and Openai told Times Happiness may not increase because the logic models are naturally worse. Instead, they can simply be more action and adventure in their answers. Because new models are not only repeating the approximate facts, but guessing the possibilities, the line between the principle and the fabricated facts can be blurred to AI. Unfortunately, some of those possibilities are completely precious from reality.

    Nevertheless, more hallucinations are open or its rivals such as Google and Anthropic are the opposite from their most advanced models. Calling AI chatbots assistants and copilots means they will be helpful, not dangerous. The lawyers have already got into trouble to use chat and not to keep in mind the quotes of the imaginary court; Who knows how many such errors have caused problems in less high-stake conditions?

    Sign up for braking news, review, opinion, top tech deals and more.

    A halight opportunities to create a problem for a user are rapidly expanding rapidly because AI systems start rolling out in classrooms, offices, hospitals and government agencies. Sophisticated AI can help prepare job applications, solve billing issues or analyze spreadsheets, but the contradiction is that more useful AI becomes, there is a less room for error.

    You cannot claim people to save time and effort, if they have to spend everything you say as double-checking. It is not that these models are not impressive. GPT O3 has performed some amazing tricks of coding and logic. It can also improve many humans in some ways. The problem is that the moment it decides whether Abraham Lincoln has hosted a podcast or the water boils at 80 ° F, the illusion of credibility.

    Until those issues are resolved, you should take any response with a teaspoon of salt from an AI model. Sometimes, the chatter is a bit like a annoying man who joins all of us in many meetings; With confidence in completely nonsense.

    You might also like

    Chatgpt hallucinations smarter spilling
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWorld Liberty Financial floats for WLFI holders USD1 Aircraft
    Next Article „Eine Krisensituation Erfordert Klare Entscheidungen”
    PineapplesUpdate
    • Website

    Related Posts

    Startups

    Advertisements are coming on Chatgpt. Here’s how they’ll work

    January 16, 2026
    Startups

    I’ve tested Gemini, ChatGPT, Copilot, and others – Lenovo has all the AI ​​assistants to beat

    January 10, 2026
    Startups

    This couple’s $1.5M holiday side hustle uses ChatGPT to grow

    December 15, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Microsoft’s new text editor is a VIM and Nano option

    May 19, 2025797 Views

    The best luxury car for buyers for the first time in 2025

    May 19, 2025724 Views

    Massives Datenleck in Cloud-Spichenn | CSO online

    May 19, 2025650 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    10,000 steps or Japanese walk? We ask experts if you should walk ahead or fast

    June 16, 20250 Views

    FIFA Club World Cup Soccer: Stream Palmirus vs. Porto lives from anywhere

    June 16, 20250 Views

    Google tests AI-operated audio overview in search results for some questions

    June 16, 20250 Views
    Our Picks

    I tried 0patch as a last resort for my Windows 10 PC – here’s how it compares to its promises

    January 20, 2026

    A PC Expert Explains Why Don’t Use Your Router’s USB Port When These Options Are Present

    January 20, 2026

    New ‘Remote Labor Index’ shows AI fails 97% of the time in freelancer tasks

    January 19, 2026

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms And Conditions
    • Disclaimer
    © 2026 PineapplesUpdate. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.