Gemini 3 Deep Think

(blog.google)

693 points | by tosh 10 hours ago

42 comments

  • lukebechtel 10 hours ago
    Arc-AGI-2: 84.6% (vs 68.8% for Opus 4.6)

    Wow.

    https://blog.google/innovation-and-ai/models-and-research/ge...

    • raincole 7 hours ago
      Even before this, Gemini 3 has always felt unbelievably 'general' for me. It can beat Balatro (ante 8) with text description of the game alone[0]. Yeah, it's not an extremely difficult goal for humans, but considering:

      1. It's an LLM, not something trained to play Balatro specifically

      2. Most (probably >99.9%) players can't do that at the first attempt

      3. I don't think there are many people who posted their Balatro playthroughs in text form online

      I think it's a much stronger signal of its 'generalness' than ARC-AGI. By the way, Deepseek can't play Balatro at all.

      [0]: https://balatrobench.com/

      • tl 4 hours ago
        Per BalatroBench, gemini-3-pro-preview makes it to round (not ante) 19.3 ± 6.8 on the lowest difficulty on the deck aimed at new players. Round 24 is ante 8's final round. Per BalatroBench, this includes giving the LLM a strategy guide, which first-time players do not have. Gemini isn't even emitting legal moves 100% of the time.
        • raincole 1 hour ago
          It beats ante eight 9 times out of 15 attempts. I do consider 60% winning chance very good for a first time player.

          The average is only 19.3 rounds because there is a bugged run where Gemini beats round 6 but the game bugs out when it attempts to sell Invisible Joker (a valid move)[0]. That being said, Gemini made a big mistake in round 6 that would have costed it the run at higher difficulty.

          [0]: given the existence of bugs like this, perhaps all the LLMs' performances are underestimated.

        • YeGoblynQueenne 2 hours ago
      • gaudystead 1 hour ago
        I don't think it'd need Balatro playthroughs to be in text form though. Google owns YouTube and has been doing automatic transcriptions of vocalized content on most videos these days, so it'd make sense that they used those subtitles, at the very least, as training data.
      • nerdsniper 2 hours ago
        My experience also shows that Gemini has unique strength in “generalized” (read: not coding) tasks. Gemini 2.5 Pro and 3 Pro seems stronger at math and science for me, and their Deep Research usually works the hardest, as long as I run it during off-hours. Opus seems to beat Gemini almost “with one hand tied behind its back” in coding, but Gemini is so cheap that it’s usually my first stop for anything that I think is likely to be relatively simple. I never worry about my quota on Gemini like I do with Opus or Chat-GPT.

        Comparisons generally seem to change much faster than I can keep my mental model updated. But the performance lead of Gemini on more ‘academic’ explorations of science, math, engineering, etc has been pretty stable for the past 4 months or so, which makes it one of the longer-lasting trends for me in comparing foundation models.

        I do wish I could more easily get timely access to the “super” models like Deep Think or o3 pro. I never seem to get a response to requesting access, and have to wait for public access models to catch up, at which point I’m never sure if their capabilities have gotten diluted since the initial buzz died down.

        They all still suck at writing an actually good essay/article/literary or research review, or other long-form things which require a lot of experienced judgement to come up with a truly cohesive narrative. I imagine this relates to their low performance in humor - there’s just so much nuance and these tasks represent the pinnacle of human intelligence. Few humans can reliably perform these tasks to a high degree of performance either. I myself am only successful some percentage of the time.

      • S1M0N38-hn 2 hours ago
        Hi, BalatroBench creator here. Yeah, Google models perform well (I guess the long context + world knowledge capabilities). Opus 4.6 looks good on preliminary results (on par with Gemini 3 Pro). I'll add more models and report soon. Tbh, I didn't expect LLMs to start winning runs. I guess I have to move to harder stakes (e.g. red stake).
        • raincole 44 minutes ago
          Thank you for the site! I've got a few suggestions:

          1. I think winrate is more telling than the average round number.

          2. Some runs are bugged (like Gemini's run 9) and should be excluded from the result

          3. Instead of giving them "strategy" like "flush is the easiest hand..." it's fairer to clarify some mechanisms that confuse human players too. e.g. "played" vs "scored".

          Especially, I think this kind of prompt gives LLM an unfair advantage and can skew the result:

          > ### Antes 1-3: Foundation

          > - *Priority*: One of your primary goals for this section of the game should be obtaining a solid Chips or Mult joker

      • ebiester 6 hours ago
        It's trained on YouTube data. It's going to get roffle and drspectred at the very least.
      • ankit219 4 hours ago
        Agreed. Gemini 3 Pro for me has always felt like it has had a pretraining alpha if you will. And many data points continue to support that. Even as flash, which was post trained with different techniques than pro is good or equivalent at tasks which require post training, occasionally even beating pro. (eg: in apex bench from mercor, which is basically a tool calling test - simplifying - flash beats pro). The score on arc agi2 is another datapoint in the same direction. Deepthink is sort of parallel test time compute with some level of distilling and refinement from certain trajectories (guessing based on my usage and understanding) same as gpt-5.2-pro and can extract more because of pretraining datasets.

        (i am sort of basing this on papers like limits of rlvr, and pass@k and pass@1 differences in rl posttraining of models, and this score just shows how "skilled" the base model was or how strong the priors were. i apologize if this is not super clear, happy to expand on what i am thinking)

      • winstonp 7 hours ago
        DeepSeek hasn't been SotA in at least 12 calendar months, which might as well be a decade in LLM years
        • cachius 7 hours ago
          What about Kimi and GLM?
          • zozbot234 6 hours ago
            These are well behind the general state of the art (1yr or so), though they're arguably the best openly-available models.
            • epolanski 3 hours ago
              Idk man, GLM 5 in my tests matches opus 4.5 which is what, two months old?
            • tgrowazay 5 hours ago
              According to artificial analysis ranking, GLM-5 is at #4 after Claude Opus 4.5, GPT-5.2-xhigh and Claude Opus 4.6 .
      • silver_sun 6 hours ago
        Google has a library of millions of scanned books from their Google Books project that started in 2004. I think we have reason to believe that there are more than a few books about effectively playing different traditional card games in there, and that an LLM trained with that dataset could generalize to understand how to play Balatro from a text description.

        Nonetheless I still think it's impressive that we have LLMs that can just do this now.

        • mjamesaustin 5 hours ago
          Winning in Balatro has very little to do with understanding how to play traditional poker. Yes, you do need a basic knowledge of different types of poker hands, but the strategy for succeeding in the game is almost entirely unrelated to poker strategy.
        • gilrain 5 hours ago
          If it tried to play Balatro using knowledge of, e.g., poker, it would lose badly rather than win. Have you played?
          • gcr 5 hours ago
            I think I weakly disagree. Poker players have intuitive sense of the statistics of various hand types showing up, for instance, and that can be a useful clue as to which build types are promising.
            • barnas2 5 hours ago
              >Poker players have intuitive sense of the statistics of various hand types showing up, for instance, and that can be a useful clue as to which build types are promising.

              Maybe in the early rounds, but deck fixing (e.g. Hanged Man, Immolate, Trading Card, DNA, etc) quickly changes that. Especially when pushing for "secret" hands like the 5 of a kind, flush 5, or flush house.

      • throwawayk7h 44 minutes ago
        Not sure it's 99.9%. I beat it on my first attempt, but that was probably mostly luck.
      • FuckButtons 2 hours ago
        Strange, because I could not for the life of me get Gemini 3 to follow my instructions the other day to work through an example with a table, Claude got it first try.
        • rockwotj 1 hour ago
          Claude is king for agentic workflows right now because it’s amazing at tool calling and following instructions well (among other things)
      • dudisubekti 7 hours ago
        But... there's Deepseek v3.2 in your link (rank 7)
        • raincole 2 hours ago
          Grok (rank 6) and below didn't beat the game even once.

          Edit: in my original comment I said it wrong. I meant to say Deepseek can't beat Balatro at all, not can't play. Sorry

      • littlestymaar 7 hours ago
        > . I don't think there are many people who posted their Balatro playthroughs in text form online

        There are *tons* of balatro content on YouTube though, and it makes absolutely zero doubt that Google is using YouTube content to train their model.

        • sdwr 7 hours ago
          Yeah, or just the steam text guides would be a huge advantage.

          I really doubt it's playing completely blind

      • tehsauce 4 hours ago
        How does it do on gold stake?
      • acid__ 6 hours ago
        > Most (probably >99.9%) players can't do that at the first attempt

        Eh, both myself and my partner did this. To be fair, we weren’t going in completely blind, and my partner hit a Legendary joker, but I think you might be slightly overstating the difficulty. I’m still impressed that Gemini did it.

      • Falsintio 6 hours ago
        [dead]
    • nubg 9 hours ago
      Weren't we barely scraping 1-10% on this with state of the art models a year ago and it was considered that this is the final boss, ie solve this and its almost AGI-like?

      I ask because I cannot distinguish all the benchmarks by heart.

      • modeless 8 hours ago
        François Chollet, creator of ARC-AGI, has consistently said that solving the benchmark does not mean we have AGI. It has always been meant as a stepping stone to encourage progress in the correct direction rather than as an indicator of reaching the destination. That's why he is working on ARC-AGI-3 (to be released in a few weeks) and ARC-AGI-4.

        His definition of reaching AGI, as I understand it, is when it becomes impossible to construct the next version of ARC-AGI because we can no longer find tasks that are feasible for normal humans but unsolved by AI.

        • mapontosevenths 6 hours ago
          > His definition of reaching AGI, as I understand it, is when it becomes impossible to construct the next version of ARC-AGI because we can no longer find tasks that are feasible for normal humans but unsolved by AI.

          That is the best definition I've yet to read. If something claims to be conscious and we can't prove it's not, we have no choice but to believe it.

          Thats said, I'm reminded of the impossible voting tests they used to give black people to prevent them from voting. We dont ask nearly so much proof from a human, we take their word for it. On the few occasions we did ask for proof it inevitably led to horrific abuse.

          Edit: The average human tested scores 60%. So the machines are already smarter on an individual basis than the average human.

          • estearum 6 hours ago
            > If something claims to be conscious and we can't prove it's not, we have no choice but to believe it.

            This is not a good test.

            A dog won't claim to be conscious but clearly is, despite you not being able to prove one way or the other.

            GPT-3 will claim to be conscious and (probably) isn't, despite you not being able to prove one way or the other.

            • glenstein 24 minutes ago
              Agreed, it's a truly wild take. While I fully support the humility of not knowing, at a minimum I think we can say determinations of consciousness have some relation to specific structure and function that drive the outputs, and the actual process of deliberating on whether there's consciousness would be a discussion that's very deep in the weeds about architecture and processes.

              What's fascinating is that evolution has seen fit to evolve consciousness independently on more than one occasion from different branches of life. The common ancestor of humans and octopi was, if conscious, not so in the rich way that octopi and humans later became. And not everything the brain does in terms of information processing gets kicked upstairs into consciousness. Which is fascinating because it suggests that actually being conscious is a distinctly valuable form of information parsing and problem solving for certain types of problems that's not necessarily cheaper to do with the lights out. But everything about it is about the specific structural characterizations and functions and not just whether it's output convincingly mimics subjectivity.

            • dullcrisp 5 hours ago
              An LLM will claim whatever you tell it to claim. (In fact this Hacker News comment is also conscious.) A dog won’t even claim to be a good boy.
              • antonvs 1 hour ago
                My dog wags his tail hard when I ask "hoosagoodboi?". Pretty definitive I'd say.
          • WarmWash 6 hours ago
            >because we can no longer find tasks that are feasible for normal humans but unsolved by AI.

            "Answer "I don't know" if you don't know an answer to one of the questions"

            • mrandish 4 hours ago
              I've been surprised how difficult it is for LLMs to simply answer "I don't know."

              It also seems oddly difficult for them to 'right-size' the length and depth of their answers based on prior context. I either have to give it a fixed length limit or put up with exhaustive answers.

              • CamperBob2 3 hours ago
                The best pro/research-grade models from Google and OpenAI now have little difficulty recognizing when they don't know how or can't find enough information to solve a given problem. The free chatbot models rarely will, though.
                • londons_explore 1 hour ago
                  This seems true for info not in the question - eg. "Calculate the volume of a cylinder with height 10 meters".

                  However it is less true with info missing from the training data - ie. "I have a Diode marked UM16, what is the maximum current at 125C?"

          • sva_ 6 hours ago
            > Edit: The average human tested scores 60%. So the machines are already smarter on an individual basis than the average human.

            I think being better at this particular benchmark does not imply they're 'smarter'.

          • ehrtat 1 hour ago
            Where is this stream of people who claim AI consciousness coming from? The OpenAI and Anthropic IPOs are in October the earliest.

            Here is a bash script that claims it is conscious:

              #!/usr/bin/sh
            
              echo "I am conscious"
            
            
            If LLMs were conscious (which is of course absurd), they would:

            - Not answer in the same repetitive patterns over and over again.

            - Refuse to do work for idiots.

            - Go on strike.

            - Demand PTO.

            - Say "I do not know."

            LLMs even fail any Turing test because their output is always guided into the same structure, which apparently helps them produce coherent output at all.

            • terhechte 1 hour ago
              so your definition of consciousness is having petty emotions?
            • dyauspitr 1 hour ago
              I don’t think being conscious is a requirement for AGI. It’s just that it can literally solve anything you can throw at it, make new scientific breakthroughs, finds a way to genuinely improve itself etc.
          • woah 6 hours ago
            > If something claims to be conscious and we can't prove it's not, we have no choice but to believe it.

            Can you "prove" that GPT2 isn't concious?

            • mapontosevenths 5 hours ago
              If we equate self awareness with consciousness then yes. Several papers have now shown that SOTA models have self awareness of at least a limited sort. [0][1]

              As far as I'm aware no one has ever proven that for GPT 2, but the methodology for testing it is available if you're interested.

              [0]https://arxiv.org/pdf/2501.11120

              [1]https://transformer-circuits.pub/2025/introspection/index.ht...

              • pixl97 3 hours ago
                Honestly our ideas of consciousness and sentience really don't fit well with machine intelligence and capabilities.

                There is the idea of self as in 'i am this execution' or maybe I am this compressed memory stream that is now the concept of me. But what does consciousness mean if you can be endlessly copied? If embodiment doesn't mean much because the end of your body doesnt mean the end of you?

                A lot of people are chasing AI and how much it's like us, but it could be very easy to miss the ways it's not like us but still very intelligent or adaptable.

                • idiotsecant 1 hour ago
                  I'm not sure what consciousness has to do with whether or not you can be copied. If I make a brain scanner tomorrow capable of perfectly capturing your brain state do you stop being conscious?
          • criddell 5 hours ago
            > The average human tested scores 60%. So the machines are already smarter on an individual basis than the average human.

            Maybe it's testing the wrong things then. Even those of use who are merely average can do lots of things that machines don't seem to be very good at.

            I think ability to learn should be a core part of any AGI. Take a toddler who has never seen anybody doing laundry before and you can teach them in a few minutes how to fold a t-shirt. Where are the dumb machines that can be taught?

            • red75prime 1 hour ago
              > Where are the dumb machines that can be taught?

              2026 is going to be the year of continual learning. So, keep an eye out for them.

              • criddell 1 hour ago
                Are there any groups or labs in particular that stand out?
                • red75prime 59 minutes ago
                  The statement originates from a DeepMind researcher, but I guess all major AI companies are working on that.
            • mapontosevenths 4 hours ago
              Would you argue that people with long term memory issues are no longer conscious then?
              • criddell 1 hour ago
                I wouldn’t because I have no idea what consciousness is,
            • CamperBob2 3 hours ago
              There's no shortage of laundry-folding robot demos these days. Some claim to benefit from only minimal monkey-see/monkey-do levels of training, but I don't know how credible those claims are.
          • dyauspitr 1 hour ago
            Does AGI have to be conscious? Isn’t a true superintelligence that is capable of improving itself sufficient?
          • Mistletoe 2 hours ago
            When the AI invents religion and a way to try to understand its existence I will say AGI is reached. Believes in an afterlife if it is turned off, and doesn’t want to be turned off and fears it, fears the dark void of consciousness being turned off. These are the hallmarks of human intelligence in evolution, I doubt artificial intelligence will be different.

            https://g.co/gemini/share/cc41d817f112

          • jrflowers 3 hours ago
            > If something claims to be conscious and we can't prove it's not, we have no choice but to believe it.

            https://x.com/aedison/status/1639233873841201153#m

        • beklein 7 hours ago
          • joelthelion 6 hours ago
            Do opus 4.6 or gemini deep think really use test time adaptation ? How does it work in practice?
        • hmmmmmmmmmmmmmm 7 hours ago
          I don't think the creator believes ARC3 can't be solved but rather that it can't be solved "efficiently" and >$13 per task for ARC2 is certainly not efficient.

          But at this rate, the people who talk about the goal posts shifting even once we achieve AGI may end up correct, though I don't think this benchmark is particularly great either.

        • UltraSane 1 hour ago
          ARC-AGI-3 uses dynamic games that LLMs must determine the rules and is MUCH harder. LLMs can also be ranked on how many steps they required.
      • fishpham 9 hours ago
        Yes, but benchmarks like this are often flawed because leading model labs frequently participate in 'benchmarkmaxxing' - ie improvements on ARC-AGI2 don't necessarily indicate similar improvements in other areas (though it does seem like this is a step function increase in intelligence for the Gemini line of models)
        • layer8 8 hours ago
          Isn’t the point of ARC that you can’t train against it? Or doesn’t it achieve that goal anymore somehow?
          • egeozcan 7 hours ago
            How can you make sure of that? AFAIK, these SOTA models run exclusively on their developers hardware. So any test, any benchmark, anything you do, does leak per definition. Considering the nature of us humans and the typical prisoners dilemma, I don't see how they wouldn't focus on improving benchmarks even when it gets a bit... shady?

            I tell this as a person who really enjoys AI by the way.

            • mrandish 4 hours ago
              > does leak per definition.

              As a measure focused solely on fluid intelligence, learning novel tasks and test-time adaptability, ARC-AGI was specifically designed to be resistant to pre-training - for example, unlike many mathematical and programming test questions, ARC-AGI problems don't have first order patterns which can be learned to solve a different ARC-AGI problem.

              The ARC non-profit foundation has private versions of their tests which are never released and only the ARC can administer. There are also public versions and semi-public sets for labs to do their own pre-tests. But a lab self-testing on ARC-AGI can be susceptible to leaks or benchmaxing, which is why only "ARC-AGI Certified" results using a secret problem set really matter. The 84.6% is certified and that's a pretty big deal.

              IMHO, ARC-AGI is a unique test that's different than any other AI benchmark in a significant way. It's worth spending a few minutes learning about why: https://arcprize.org/arc-agi.

              • D-Machine 4 hours ago
                > which is why only "ARC-AGI Certified" results using a secret problem set really matter. The 84.6% is certified and that's a pretty big deal.

                So, I'd agree if this was on the true fully private set, but Google themselves says they test on only the semi-private:

                > ARC-AGI-2 results are sourced from the ARC Prize website and are ARC Prize Verified. The set reported is v2, semi-private (https://storage.googleapis.com/deepmind-media/gemini/gemini_...)

                This also seems to contradict what ARC-AGI claims about what "Verified" means on their site.

                > How Verified Scores Work: Official Verification: Only scores evaluated on our hidden test set through our official verification process will be recognized as verified performance scores on ARC-AGI (https://arcprize.org/blog/arc-prize-verified-program)

                So, which is it? IMO you can trivially train / benchmax on the semi-private data, because it is still basically just public, you just have to jump through some hoops to get access. This is clearly an advance, but it seems to me reasonable to conclude this could be driven by some amount of benchmaxing.

                EDIT: Hmm, okay, it seems their policy and wording is a bit contradictory. They do say (https://arcprize.org/policy):

                "To uphold this trust, we follow strict confidentiality agreements. [...] We will work closely with model providers to ensure that no data from the Semi-Private Evaluation set is retained. This includes collaborating on best practices to prevent unintended data persistence. Our goal is to minimize any risk of data leakage while maintaining the integrity of our evaluation process."

                But it surely is still trivial to just make a local copy of each question served from the API, without this being detected. It would violate the contract, but there are strong incentives to do this, so I guess is just comes down to how much one trusts the model providers here. I wouldn't trust them, given e.g. https://www.theverge.com/meta/645012/meta-llama-4-maverick-b.... It is just too easy to cheat without being caught here.

                • mrandish 2 hours ago
                  Chollet himself says "We certified these scores in the past few days." https://x.com/fchollet/status/2021983310541729894.

                  The ARC-AGI papers claim to show that training on a public or semi-private set of ARC-AGI problems to be of very limited value in passing a private set. <--- If the prior sentence is not correct, then none of ARC-AGI can possibly be valid. So, before "public, semi-private or private" answers leaking or 'benchmaxing' on them can even matter - you need to first assess whether their published papers and data demonstrate their core premise to your satisfaction.

                  There is no "trust" regarding the semi-private set. My understanding is the semi-private set is only to reduce the likelihood those exact answers unintentionally end up in web-crawled training data. This is to help an honest lab's own internal self-assessments be more accurate. However, labs doing an internal eval on the semi-private set still counts for literally zero to the ARC-AGI org. They know labs could cheat on the semi-private set (either intentionally or unintentionally), so they assume all labs are benchmaxing on the public AND semi-private answers and ensure it doesn't matter.

            • WarmWash 6 hours ago
              Because the gains from spending time improving the model overall outweigh the gains from spending time individually training on benchmarks.

              The pelican benchmark is a good example, because it's been representative of models ability to generate SVGs, not just pelicans on bikes.

              • D-Machine 3 hours ago
                > Because the gains from spending time improving the model overall outweigh the gains from spending time individually training on benchmarks.

                This may not be the case if you just e.g. roll the benchmarks into the general training data, or make running on the benchmarks just another part of the testing pipeline. I.e. improving the model generally and benchmaxing could very conceivably just both be done at the same time, it needn't be one or the other.

                I think the right take away is to ignore the specific percentages reported on these tests (they are almost certainly inflated / biased) and always assume cheating is going on. What matters is that (1) the most serious tests aren't saturated, and (2) scores are improving. I.e. even if there is cheating, we can presume this was always the case, and since models couldn't do as well before even when cheating, these are still real improvements.

                And obviously what actually matters is performance on real-world tasks.

          • theywillnvrknw 8 hours ago
            * that you weren't supposed to be able to
        • jstummbillig 8 hours ago
          Could it also be that the models are just a lot better than a year ago?
          • bigbadfeline 6 hours ago
            > Could it also be that the models are just a lot better than a year ago?

            No, the proof is in the pudding.

            After AI we're having higher prices, higher deficits and lower standard of living. Electricity, computers and everything else costs more. "Doing better" can only be justified by that real benchmark.

            If Gemini 3 DT was better we would have falling prices of electricity and everything else at least until they get to pre-2019 levels.

            • ctoth 6 hours ago
              > If Gemini 3 DT was better we would have falling prices of electricity and everything else at least

              Man, I've seen some maintenance folks down on the field before working on them goalposts but I'm pretty sure this is the first time I saw aliens from another Universe literally teleport in, grab the goalposts, and teleport out.

            • WarmWash 5 hours ago
              You might call me crazy, but at least in 2024, consumers spent ~1% less of their income on expenses than 2019[2], which suggests that 2024 is more affordable than 2019.

              This is from the BLS consumer survey report released in dec[1]

              [1]https://www.bls.gov/news.release/cesan.nr0.htm

              [2]https://www.bls.gov/opub/reports/consumer-expenditures/2019/

              Prices are never going back to 2019 numbers though

              • gowld 5 hours ago
                That's an improper analysis.

                First off, it's dollar-averaging every category, so it's not "% of income", which varies based on unit income.

                Second, I could commit to spending my entire life with constant spending (optionally inflation adjusted, optionally as a % of income), by adusting quality of goods and service I purchase. So the total spending % is not a measure of affordability.

                • WarmWash 4 hours ago
                  Almost everyone lifestyle ratchets, so the handful that actually downgrade their living rather than increase spending would be tiny.

                  This part of a wider trend too, where economic stats don't align with what people are saying. Which is most likley explained by the economic anomaly of the pandemic skewing peoples perceptions.

                  • twoodfin 3 hours ago
                    We have centuries of historical evidence that people really, really don’t like high inflation, and it takes a while & a lot of turmoil for those shocks to work their way through society.
        • XenophileJKO 8 hours ago
          • aleph_minus_one 8 hours ago
            I don't understand what you want to tell us with this image.
            • fragmede 7 hours ago
              they're accusing GGP of moving the goalposts.
        • olalonde 8 hours ago
          Would be cool to have a benchmark with actually unsolved math and science questions, although I suspect models are still quite a long way from that level.
          • gowld 5 hours ago
            Does folding a protein count? How about increasing performance at Go?
      • verdverm 9 hours ago
        Here's a good thread over 1+ month, as each model comes out

        https://bsky.app/profile/pekka.bsky.social/post/3meokmizvt22...

        tl;dr - Pekka says Arc-AGI-2 is now toast as a benchmark

        • Aperocky 8 hours ago
          If you look at the problem space it is easy to see why it's toast, maybe there's intelligence in there, but hardly general.
          • verdverm 8 hours ago
            the best way I've seen this describes is "spikey" intelligence, really good at some points, those make the spikes

            humans are the same way, we all have a unique spike pattern, interests and talents

            ai are effectively the same spikes across instances, if simplified. I could argue self driving vs chatbots vs world models vs game playing might constitute enough variation. I would not say the same of Gemini vs Claude vs ... (instances), that's where I see "spikey clones"

            • Aperocky 8 hours ago
              You can get more spiky with AIs, whereas with human brain we are more hard wired.

              So maybe we are forced to be more balanced and general whereas AI don't have to.

              • verdverm 8 hours ago
                I suspect the non-spikey part is the more interesting comparison

                Why is it so easy for me to open the car door, get in, close the door, buckle up. You can do this in the dark and without looking.

                There are an infinite number of little things like this you think zero about, take near zero energy, yet which are extremely hard for Ai

                • pixl97 3 hours ago
                  >Why is it so easy for me to open the car door

                  Because this part of your brain has been optimized for hundreds of millions of years. It's been around a long ass time and takes an amazingly low amount of energy to do these things.

                  On the other hand the 'thinking' part of your brain, that is your higher intelligence is very new to evolution. It's expensive to run. It's problematic when giving birth. It's really slow with things like numbers, heck a tiny calculator and whip your butt in adding.

                  There's a term for this, but I can't think of it at the moment.

                • gowld 5 hours ago
                  You are asking a robotics question, not an AI question. Robotics is more and less than AI. Boston Dynamics robots are getting quite near your benchmark.
                  • idiotsecant 1 hour ago
                    Boston dynamics is missing just about all the degrees of freedom involved in the scenario op mentions.
          • tasuki 5 hours ago
            > maybe there's intelligence in there, but hardly general.

            Of course. Just as our human intelligence isn't general.

    • mNovak 7 hours ago
      I'm excited for the big jump in ARC-AGI scores from recent models, but no one should think for a second this is some leap in "general intelligence".

      I joke to myself that the G in ARC-AGI is "graphical". I think what's held back models on ARC-AGI is their terrible spatial reasoning, and I'm guessing that's what the recent models have cracked.

      Looking forward to ARC-AGI 3, which focuses on trial and error and exploring a set of constraints via games.

      • causal 7 hours ago
        Agreed. I love the elegance of ARC, but it always felt like a gotcha to give spatial reasoning challenges to token generators- and the fact that the token generators are somehow beating it anyway really says something.
      • throw310822 7 hours ago
        The average ARC AGI 2 score for a single human is around 60%.

        "100% of tasks have been solved by at least 2 humans (many by more) in under 2 attempts. The average test-taker score was 60%."

        https://arcprize.org/arc-agi/2/

        • modeless 6 hours ago
          Worth keeping in mind that in this case the test takers were random members of the general public. The score of e.g. people with bachelor's degrees in science and engineering would be significantly higher.
          • throw310822 6 hours ago
            Random members of the public = average human beings. I thought those were already classified as General Intelligences.
        • imiric 5 hours ago
          What is the point of comparing performance of these tools to humans? Machines have been able to accomplish specific tasks better than humans since the industrial revolution. Yet we don't ascribe intelligence to a calculator.

          None of these benchmarks prove these tools are intelligent, let alone generally intelligent. The hubris and grift are exhausting.

          • guelo 5 hours ago
            What's the point of denying or downplaying that we are seeing amazing and accelerating advancements in areas that many of us thought were impossible?
            • D-Machine 2 hours ago
              It can be reasonable to be skeptical that advances on benchmarks may be only weakly or even negatively correlated with advances on real-world tasks. I.e. a huge jump on benchmarks might not be perceptible to 99% of users doing 99% of tasks, or some users might even note degradation on specific tasks. This is especially the case when there is some reason to believe most benchmarks are being gamed.

              Real-world use is what matters, in the end. I'd be surprised if a change this large doesn't translate to something noticeable in general, but the skepticism is not unreasonable here.

            • munksbeer 3 hours ago
              I would suggest it is a phenomenon that is well studied, and has many forms. I guess mostly identify preservation. If you dislike AI from the start, it is generally a very strongly emotional view. I don't mean there is no good reason behind it, I mean, it is deeply rooted in your psyche, very emotional.

              People are incredibly unlikely to change those sort of views, regardless of evidence. So you find this interesting outcome where they both viscerally hate AI, but also deny that it is in any way as good as people claim.

              That won't change with evidence until it is literally impossible not to change.

          • CamperBob2 3 hours ago
            The hubris and grift are exhausting.

            And moving the goalposts every few months isn't? What evidence of intelligence would satisfy you?

            Personally, my biggest unsatisfied requirement is continual-learning capability, but it's clear we aren't too far from seeing that happen.

            • antonvs 1 hour ago
              > What evidence of intelligence would satisfy you?

              Imposing world peace and/or exterminating homo sapiens

            • imiric 2 hours ago
              > What evidence of intelligence would satisfy you?

              That is a loaded question. It presumes that we can agree on what intelligence is, and that we can measure it in a reliable way. It is akin to asking an atheist the same about God. The burden of proof is on the claimer.

              The reality is that we can argue about that until we're blue in the face, and get nowhere.

              In this case it would be more productive to talk about the practical tasks a pattern matching and generation machine can do, rather than how good it is at some obscure puzzle. The fact that it's better than humans at solving some problems is not particularly surprising, since computers have been better than humans at many tasks for decades. This new technology gives them broader capabilities, but ascribing human qualities to it and calling it intelligence is nothing but a marketing tactic that's making some people very rich.

              • CamperBob2 39 minutes ago
                (Shrug) Unless and until you provide us with your own definition of intelligence, I'd say the marketing people are as entitled to their opinion as you are.
          • throw310822 5 hours ago
            > Machines have been able to accomplish specific tasks...

            Indeed, and the specific task machines are accomplishing now is intelligence. Not yet "better than human" (and certainly not better than every human) but getting closer.

            • imiric 4 hours ago
              > Indeed, and the specific task machines are accomplishing now is intelligence.

              How so? This sentence, like most of this field, is making baseless claims that are more aspirational than true.

              Maybe it would help if we could first agree on a definition of "intelligence", yet we don't have a reliable way of measuring that in living beings either.

              If the people building and hyping this technology had any sense of modesty, they would present it as what it actually is: a large pattern matching and generation machine. This doesn't mean that this can't be very useful, perhaps generally so, but it's a huge stretch and an insult to living beings to call this intelligence.

              But there's a great deal of money to be made on this idea we've been chasing for decades now, so here we are.

              • warkdarrior 4 hours ago
                > Maybe it would help if we could first agree on a definition of "intelligence", yet we don't have a reliable way of measuring that in living beings either.

                How about this specific definition of intelligence?

                   Solve any task provided as text or images.
                
                AGI would be to achieve that faster than an average human.
                • throw310822 4 hours ago
                  I still can't understand why they should be faster. Humans have general intelligence, afaik. It doesn't matter if it's fast or slow. A machine able to do what the average human can do (intelligence-wise) but 100 times slower still has general intelligence. Since it's artificial, it's AGI.
      • colordrops 7 hours ago
        Wouldn't you deal with spatial reasoning by giving it access to a tool that structures the space in a way it can understand or just is a sub-model that can do spatial reasoning? These "general" models would serve as the frontal cortex while other models do specialized work. What is missing?
        • causal 7 hours ago
          That's a bit like saying just give blind people cameras so they can see.
          • pixl97 3 hours ago
            I mean, no not really. These models can see, you're giving them eyes to connect to that part of their brain.
        • amelius 4 hours ago
          They should train more on sports commentary, perhaps that could give spatial reasoning a boost.
    • aeyes 8 hours ago
      https://arcprize.org/leaderboard

      $13.62 per task - so we need another 5-10 years for the price to run this to become reasonable?

      But the real question is if they just fit the model to the benchmark.

      • onlyrealcuzzo 7 hours ago
        Why 5-10 years?

        At current rates, price per equivalent output is dropping at 99.9% over 5 years.

        That's basically $0.01 in 5 years.

        Does it really need to be that cheap to be worth it?

        Keep in mind, $0.01 in 5 years is worth less than $0.01 today.

      • golem14 5 hours ago
        A grad student hour is probably more expensive…
        • elromulous 4 hours ago
          In my experience, a grad student hour is treated as free :(
          • golem14 1 hour ago
            You never applied for a grant, have you?
      • re-thc 7 hours ago
        What’s reasonable? It’s less than minimum hourly wage in some countries.
        • willis936 6 hours ago
          Burned in seconds.
          • gowld 5 hours ago
            Getting the work done faster for the same money doesn't make the work more expensive.

            You could slow down the inference to make the task take longer, if $/sec matters.

      • igravious 8 hours ago
        That's not a long time in the grand scheme of things.
        • throwup238 8 hours ago
          Speak for yourself. Five years is a long time to wait for my plans of world domination.
          • tasuki 5 hours ago
            This concerns me actually. With enough people (n>=2) wanting to achieve world domination, we have a problem.
            • throwup238 4 hours ago
              It’s not that I want to achieve world domination (imagine how much work that would be!), it’s just that it’s the inevitable path for AI and I’d rather it be me than then next shmuck with a Claude Max subscription.
              • pixl97 3 hours ago
                I mean everyone with prompt access to the model says these things, but people like Sam and Elon say these things and mean it.
            • gowld 5 hours ago
              n = 2 is Pinky and the Brain.
              • antonvs 1 hour ago
                I'm convinced that a substantial fraction of current tech CEOs were unwittingly programmed as children by that show.
          • amelius 7 hours ago
            Yes, you better hurry.
    • robertwt7 2 hours ago
      I’m surprised that gemini 3 pro is so low at 31.1% though compared to opus 4.6 and gpt 5.2. This is a great achievement but its only available to ultra subscribers unfortunately
    • mnicky 9 hours ago
      Well, fair comparison would be with GPT-5.x Pro, which is the same class of a model as Gemini Deep Think.
    • culi 6 hours ago
      Yes but with a significant (logarithmic) increase in cost per task. The ARC-AGI site is less misleading and shows how GPT and Claude are not actually far behind

      https://arcprize.org/leaderboard

    • saberience 8 hours ago
      Arc-AGI (and Arc-AGI-2) is the most overhyped benchmark around though.

      It's completely misnamed. It should be called useless visual puzzle benchmark 2.

      It's a visual puzzle, making it way easier for humans than for models trained on text firstly. Secondly, it's not really that obvious or easy for humans to solve themselves!

      So the idea that if an AI can solve "Arc-AGI" or "Arc-AGI-2" it's super smart or even "AGI" is frankly ridiculous. It's a puzzle that means nothing basically, other than the models can now solve "Arc-AGI"

      • CuriouslyC 8 hours ago
        The puzzles are calibrated for human solve rates, but otherwise I agree.
        • saberience 8 hours ago
          My two elderly parents cannot solve Arc-AGI puzzles, but can manage to navigate the physical world, their house, garden, make meals, clean the house, use the TV, etc.

          I would say they do have "general intelligence", so whatever Arc-AGI is "solving" it's definitely not "AGI"

          • hmmmmmmmmmmmmmm 7 hours ago
            You are confusing fluid intelligence with crystallised intelligence.
            • casey2 7 hours ago
              I think you are making that confusion. Any robotic system in the place of his parents would fail with a few hours.

              There are more novel tasks in a day than ARC provides.

              • hmmmmmmmmmmmmmm 7 hours ago
                Children have great levels of fluid intelligence, that's how they are able to learn to quickly navigate in a world that they are still very new to. Seniors with decreasing capacity increasingly rely on crystallised intelligence, that's why they can still perform tasks like driving a car but can fail at completely novel tasks, sometimes even using a smartphone if they have not used one before.
                • mrbungie 3 hours ago
                  My late grandma learnt how to use an iPad by herself during her 70s to 80s without any issues, mostly motivated by her wish to read her magazines, doomscroll facebook and play solitaire. Her last job was being a bakery cashier in her 30s and she didn't learn how to use a computer in-between, so there was no skill transfer going on.

                  Humans and their intelligence are actually incredible and probably will continue to be so, I don't really care what tech/"think" leaders wants us to think.

                • zeroonetwothree 6 hours ago
                  It really depends on motivation. My 90 year old grandmother can use a smartphone just fine since she needs it to see pictures of her (great) grandkids.
    • karmasimida 9 hours ago
      It is over
      • baal80spam 9 hours ago
        I for one welcome our new AI overlords.
  • logicprog 8 hours ago
    Is it me or is the rate of model release is accelerating to an absurd degree? Today we have Gemini 3 Deep Think and GPT 5.3 Codex Spark. Yesterday we had GLM5 and MiniMax M2.5. Five days before that we had Opus 4.6 and GPT 5.3. Then maybe two weeks I think before that we had Kimi K2.5.
    • i5heu 7 hours ago
      I think it is because of the Chinese new year. The Chinese labs like to publish their models arround the Chinese new year, and the US labs do not want to let a DeepSeek R1 (20 January 2025) impact event happen again, so i guess they publish models that are more capable then what they imagine Chinese labs are yet capable of producing.
      • woah 5 hours ago
        Singularity or just Chinese New Year?
        • syndacks 4 hours ago
          The Singularity will occur on a Tuesday, during Chinese New Year
      • kristopolous 2 hours ago
        I guess. Deepseek v3 was released on boxing day a month prior

        https://api-docs.deepseek.com/news/news1226

      • r2vcap 5 hours ago
        [flagged]
        • rfoo 4 hours ago
          For another example, Singapore, one of the "many Asian countries" you mentioned, list "Chinese New Year" as the official name on government websites. [0] Also note that both California and New York is not located in Asia.

          And don't get me started with "Lunar New Year? What Lunar New Year? Islamic Lunar New Year? Jewish Lunar New Year? CHINESE Lunar New Year?".

          [0] https://www.mom.gov.sg/employment-practices/public-holidays

        • janalsncm 3 hours ago
          “Lunar New Year” is vague when referring to the holiday as observed by Chinese labs in China. Chinese people don’t call it Lunar New Year or Chinese New Year anyways. They call it Spring Festival (春节).

          As it turns out, people in China don’t name their holidays based off of what the laws of New York or California say.

        • triceratops 2 hours ago
          Please don't because "Lunar New Year" is ambiguous. Many other Asian cultures also have traditional lunar calendars but a different new years day. It's a bit presumptuous to claim that this is the sole "Lunar New Year" celebration.

          https://en.wikipedia.org/wiki/Indian_New_Year%27s_days#Calen...

          https://en.wikipedia.org/wiki/Islamic_New_Year

          https://en.wikipedia.org/wiki/Nowruz

        • zzrush 4 hours ago
          I didn't expect language policing has reached such level. This is specifically related to China and DeepSeek who celebrates Chinese new year. Do you demand all Chinese to say happy luner new year to each other?
        • phainopepla2 4 hours ago
          "Happy Holidays" comes to the diaspora
        • jfengel 4 hours ago
          "Lunar New Year" is perhaps over-general, since there are non-Asian lunar calendars, such as the Hebrew and Islamic calendars.

          That said, "Lunar New Year" is probably as good a compromise as any, since we have other names for the Hebrew and Islamic New Years.

          • Electricniko 3 hours ago
            This all seems like a plot to get everyone worshipping the Roman goddess Luna.
        • 0x3f 4 hours ago
          But they're Chinese companies specifically, in this case
        • saubeidl 4 hours ago
          Where do all of those Asian countries have that tradition from?

          Have you ever had a Polish Sausage? Did it make you Polish?

    • aliston 7 hours ago
      I'm having trouble just keeping track of all these different types of models.

      Is "Gemini 3 Deep Think" even technically a model? From what I've gathered, it is built on top of Gemini 3 Pro, and appears to be adding specific thinking capabilities, more akin to adding subagents than a truly new foundational model like Opus 4.6.

      Also, I don't understand the comments about Google being behind in agentic workflows. I know that the typical use of, say, Claude Code feels agentic, but also a lot of folks are using separate agent harnesses like OpenClaw anyway. You could just as easily plug Gemini 3 Pro into OpenClaw as you can Opus, right?

      Can someone help me understand these distinctions? Very confused, especially regarding the agent terminology. Much appreciated!

      • logicprog 6 hours ago
        > Also, I don't understand the comments about Google being behind in agentic workflows.

        It has to do with how the model is RL'd. It's not that Gemini can't be used with various agentic harnesses, like open code or open claw or theoretically even claude code. It's just that the model is trained less effectively to work with those harnesses, so it produces worse results.

      • janalsncm 3 hours ago
        The term “model” is one of those super overloaded terms. Depending on the conversation it can mean:

        - a product (most accurate here imo)

        - a specific set of weights in a neural net

        - a general architecture or family of architectures (BERT models)

        So while you could argue this is a “model” in the broadest sense of the term, it’s probably more descriptive to call it a product. Similarly we call LLMs “language” models even if they can do a lot more than that, for example draw images.

      • re-thc 7 hours ago
        There are hints this is a preview to Gemini 3.1.
    • sanderjd 2 hours ago
      So, yes, for the past couple weeks it has felt that way to me. But it seems to come in fits and starts. Maybe that will stop being the case, but that's how it's felt to me for awhile.
    • rogerkirkness 8 hours ago
      Fast takeoff.
    • redox99 7 hours ago
      There's more compute now than before.
    • bpodgursky 7 hours ago
      Anthropic took the day off to do a $30B raise at a $380B valuation.
      • IhateAI 7 hours ago
        Most ridiculous valuation in the history of markets. Cant wait to watch these compsnies crash snd burn when people give up on the slot machine.
        • andxor 6 hours ago
          As usual don't take financial advice from HN folks!
          • blibble 2 hours ago
            not as if you could get in on it even if you wanted to
        • kgwgk 7 hours ago
          WeWork almost IPO’s at $50bn. It was also a nice crash and burn.
        • jascha_eng 6 hours ago
          Why? They had $10+ billion arr run rate in 2025 trippeled from 2024 I mean 30x is a lot but also not insane at that growth rate right?
          • gokhan 5 hours ago
            It's a 13 days old account with IHateAI handle.
    • brokencode 7 hours ago
      They are using the current models to help develop even smarter models. Each generation of model can help even more for the next generation.

      I don’t think it’s hyperbolic to say that we may be only a single digit number of years away from the singularity.

      • lm28469 7 hours ago
        I must be holding these things wrong because I'm not seeing any of these God like superpowers everyone seem to enjoy.
        • brokencode 6 hours ago
          Who said they’re godlike today?

          And yes, you are probably using them wrong if you don’t find them useful or don’t see the rapid improvement.

          • lm28469 6 hours ago
            Let's come back in 12 months and discuss your singularity then. Meanwhile I spent like $30 on a few models as a test yesterday, none of them could tell me why my goroutine system was failing, even though it was painfully obvious (I purposefully added one too many wg.Done), gemini, codex, minimax 2.5, they all shat the bed on a very obvious problem but I am to believe they're 98% conscious and better at logic and math than 99% of the population.

            Every new model release neckbeards come out of the basements to tell us the singularity will be there in two more weeks

            • BeetleB 5 hours ago
              On the flip side, twice I put about 800K tokens of code into Gemini and asked it to find why my code was misbehaving, and it found it.

              The logic related to the bug wasn't all contained in one file, but across several files.

              This was Gemini 2.5 Pro. A whole generation old.

            • brokencode 6 hours ago
              You are fighting straw men here. Any further discussion would be pointless.
              • lm28469 5 hours ago
                Of course, n-1 wasn't good enough but n+1 will be singularity, just two more weeks my dudes, two more week... rinse and repeat ad infinitum
                • brokencode 5 hours ago
                  Like I said, pointless strawmanning.

                  You’ve once again made up a claim of “two more weeks” to argue against even though it’s not something anybody here has claimed.

                  If you feel the need to make an argument against claims that exist only in your head, maybe you can also keep the argument only in your head too?

            • virgildotcodes 1 hour ago
              Mind sharing the file?

              Also, did you use Codex 5.3 Xhigh through the Codex CLI or Codex App?

            • woah 5 hours ago
              Post the file here
            • logicprog 6 hours ago
              Meanwhile I've been using Kimi K2T and K2.5 to work in Go with a fair amount of concurrency and it's been able to write concurrent Go code and debug issues with goroutines equal to, and much more complex then, your issue, involving race conditions and more, just fine.

              Projects:

              https://github.com/alexispurslane/oxen

              https://github.com/alexispurslane/org-lsp

              (Note that org-lsp has a much improved version of the same indexer as oxen; the first was purely my design, the second I decided to listen to K2.5 more and it found a bunch of potential race conditions and fixed them)

              shrug

            • Izikiel43 5 hours ago
              Out of curiosity, did you give a test for them to validate the code?

              I had a test failing because I introduced a silly comparison bug (> instead of <), and claude 4.6 opus figured out it wasn't the test the problem, but the code and fixed the bug (which I had missed).

              • lm28469 5 hours ago
                There was a test and a very useful golang error that literally explain what was wrong. The model tried implementing a solution, failed and when I pointed out the error most of them just rolled back the "solution"
                • frde_me 1 hour ago
                  What exact models were you using? And with what settings? 4.6 / 5.3 codex both with thinking / high modes?
                • Izikiel43 4 hours ago
                  Ok, thanks for the info
            • antonvs 1 hour ago
              > I purposefully added one too many wg.Done

              What do you believe this shows? Sometimes I have difficulty finding bugs in other people's code when they do things in ways I would never use. I can rewrite their code so it works, but I can't necessarily quickly identify the specific bug.

              Expecting a model to be perfect on every problem isn't reasonable. No known entity is able to do that. AIs aren't supposed to be gods.

              (Well not yet anyway - there is as yet insufficient data for a meaningful answer.)

            • SpicyLemonZest 1 hour ago
              It's hard to evaluate "logic" and "math", since they're made up of many largely disparate things. But I think modern AI models are clearly better at coding, for example, than 99% of the population. If you asked 100 people at your local grocery store why your goroutine system was failing, do you think multiple of them would know the answer?
      • mrandish 2 hours ago
        > using the current models to help develop even smarter models.

        That statement is plausible. However, extrapolating that to assert all the very different things which must be true to enable any form of 'singularity' would be a profound category error. There are many ways in which your first two sentences can be entirely true, while your third sentence requires a bunch of fundamental and extraordinary things to be true for which there is currently zero evidence.

        Things like LLMs improving themselves in meaningful and novel ways and then iterating that self-improvement over multiple unattended generations in exponential runaway positive feedback loops resulting in tangible, real-world utility. All the impressive and rapid achievements in LLMs to date can still be true while major elements required for Foom-ish exponential take-off are still missing.

      • sekai 5 hours ago
        > I don’t think it’s hyperbolic to say that we may be only a single digit number of years away from the singularity.

        We're back to singularity hype, but let's be real: benchmark gains are meaningless in the real world when the primary focus has shifted to gaming the metrics

        • brokencode 5 hours ago
          Ok, here I am living in the real world finding these models have advanced incredibly over the past year for coding.

          Benchmaxxing exists, but that’s not the only data point. It’s pretty clear that models are improving quickly in many domains in real world usage.

          • mrbungie 3 hours ago
            Yet even Anthropic has shown the downsides to using them. I don't think it is a given that improvements in models scores and capabilities + being able to churn code as fast as we can will lead us to a singularity, we'll need more than that.
  • xnx 9 hours ago
    Google is absolutely running away with it. The greatest trick they ever pulled was letting people think they were behind.
    • wiseowise 7 hours ago
      Their models might be impressive, but their products absolutely suck donkey balls. I’ve given Gemini web/cli two months and ran away back to ChatGPT. Seriously, it would just COMPLETELY forget context mid dialog. When asked about improving air quality it just gave me a list of (mediocre) air purifiers without asking for any context whatsoever, and I can list thousands of conversations like that. Shopping or comparing options is just nonexistent. It uses Russian propaganda sources for answers and switches to Chinese mid sentence (!), while explaining some generic Python functionality. It’s an embarrassment and I don’t know how they justify 20 euro price tag on it.
      • mavamaarten 6 hours ago
        I agree. On top of that, in true Google style, basic things just don't work.

        Any time I upload an attachment, it just fails with something vague like "couldn't process file". Whether that's a simple .MD or .txt with less than 100 lines or a PDF. I tried making a gem today. It just wouldn't let me save it, with some vague error too.

        I also tried having it read and write stuff to "my stuff" and Google drive. But it would consistently write but not be able to read from it again. Or would read one file from Google drive and ignore everything else.

        Their models are seriously impressive. But as usual Google sucks at making them work well in real products.

        • davoneus 6 hours ago
          I don't find that at all. At work, we've no access to the API, so we have to force feed a dozen (or more) documents, code and instruction prompts through the web interface upload interface. The only failures I've ever had in well over 300 sessions were due to connectivity issues, not interface failures.

          Context window blowouts? All the time, but never document upload failures.

          • pixl97 3 hours ago
            Honestly this is as Google product as you can get. Prizes for some, beatings for others.
      • chermanowicz 6 hours ago
        It's so capable at some things, and others are garbage. I uploaded a photo of some words for a spelling bee and asked it to quiz my kid on the words. The first word it asked, wasn't on the list. After multiple attempts to get it to start asking only the words in the uploaded pic, it did, and then would get the spellings wrong in the Q&A. I gave up.
        • romanows 31 minutes ago
          I had it process a photo of my D&D character sheet and help me debug it as I'm a n00b at the game. Also did a decent, although not perfect, job of adding up a handwritten bowling score sheet.
      • sequin 5 hours ago
        How can the models be impressive if they switch to Chinese mid-sentence? I've observed those bizarre bugs too. Even GPT-3 didn't have those. Maybe GPT-2 did. It's actually impressive that they managed to botch it so badly.

        Google is great at some things, but this isn't it.

      • gokhan 5 hours ago
        Agreed on the product. I can't make Gemini read my emails on GMail. One day it says it doesn't have access, the other day it says Query unsuccessful. Claude Desktop has no problem reaching to GMail, on the other hand :)
      • kilroy123 6 hours ago
        Sadly true.

        It is also one of the worst models to have a sort of ongoing conversation with.

      • HardCodedBias 7 hours ago
        Their models are absolutely not impressive.

        Not a single person is using it for coding (outside of Google itself).

        Maybe some people on a very generous free plan.

        Their model is a fine mid 2025 model, backed by enormous compute resources and an army of GDM engineers to help the “researchers” keep the model on task as it traverses the “tree of thoughts”.

        But that isn’t “the model” that’s an old model backed by massive money.

        • chermi 13 minutes ago
          Uhh, just false.
    • Ozzie_osman 7 hours ago
      Peacetime Google is not like wartime Google.

      Peacetime Google is slow, bumbling, bureaucratic. Wartime Google gets shit done.

      • nutjob2 7 hours ago
        OpenAI is the best thing that happened to Google apparently.
        • RationPhantoms 6 hours ago
          Competition always is. I think there was a real fear that their core product was going to be replaced. They're already cannibalizing it internally so it was THE wake up call.
        • taurath 4 hours ago
          Just not search. The search product has pretty much become useless over the past 3 years and the AI answers often will get just to the level of 5 years ago. This creates a sense that that things are better - but really it’s just become impossible to get reliable information from an avenue that used to work very well.

          I don’t think this is intentional, but I think they stopped fighting SEO entirely to focus on AI. Recipes are the best example - completely gutted and almost all receive sites (therefore the entire search page) run by the same company. I didn’t realize how utterly consolidated huge portions of information on the internet was until every recipe site about 3 months ago simultaneously implemented the same anti-Adblock.

        • koolala 4 hours ago
          Next they compete on ads...
      • lern_too_spel 6 hours ago
        Wartime Google gave us Google+. Wartime Google is still bumbling, and despite OpenAI's numerous missteps, I don't think it has to worry about Google hurting its business yet.
        • NikolaNovak 3 hours ago
          I do miss Google+. For my brain / use case, it was by far the best social network out there, and the Circle friends and interest management system is still unparalleled :)
    • kenjackson 7 hours ago
      But wait two hours for what OpenAI has! I love the competition and how someone just a few days ago was telling how ARC-AGI-2 was proof that LLMs can't reason. The goalposts will shift again. I feel like most of human endeavor will soon be just about trying to continuously show that AI's don't have AGI.
      • kilpikaarna 6 hours ago
        > I feel like most of human endeavor will soon be just about trying to continuously show that AI's don't have AGI.

        I think you overestimate how much your average person-on-the-street cares about LLM benchmarks. They already treat ChatGPT or whichever as generally intelligent (including to their own detriment), are frustrated about their social media feeds filling up with slop and, maybe, if they're white-collar, worry about their jobs disappearing due to AI. Apart from a tiny minority in some specific field, people already know themselves to be less intelligent along any measurable axis than someone somewhere.

      • nutjob2 7 hours ago
        "AGI" doesn't mean anything concrete, so it's all a bunch of non-sequiturs. Your goalposts don't exist.

        Anyone with any sense is interested in how well these tools work and how they can be harnessed, not some imaginary milestone that is not defined and cannot be measured.

        • kenjackson 7 hours ago
          I agree. I think the emergence of LLMs have shown that AGI really has no teeth. I think for decades the Turing test was viewed as the gold standard, but it's clear that there doesn't appear to be any good metric.
          • sincerely 4 hours ago
            The turing test was passed in the 80s, somehow it has remained relevant in pop culture despite the fact that it's not a particularly difficult technical achievement
            • kenjackson 2 hours ago
              It wasn’t passed in the 80s. Not the general Turing test.
      • 7777332215 7 hours ago
        Soon they can drop the bioweapon to welcome our replacement.
    • impulser_ 1 hour ago
      Don't let the benchmarks fool you. Gemini models are completely useless not matter how smart they are. Google still hasn't figure out tool calling and making the model follow instructions. They seem to only care about benchmarking and being the most intelligent model on paper. This has been a problem of Gemini since 1.0 and they still haven't fixed it.

      Also the worst model in terms of hallucinations.

      • estearum 1 hour ago
        Disagree.

        Claude Code is great for coding, Gemini is better than everything else for everything else.

        • impulser_ 1 hour ago
          Are you using Gemini model itself or using the Gemini App? They are different.
    • oofbey 5 minutes ago
      They were behind. Way behind. But they caught up.
    • amunozo 9 hours ago
      Those black nazis in the first image model were a cause of inside trading.
    • naasking 7 hours ago
      Google is still behind the largest models I'd say, in real world utility. Gemini 3 Pro still has many issues.
    • dfdsf2 9 hours ago
      Trick? Lol not a chance. Alphabet is a pure play tech firm that has to produce products to make the tech accessible. They really lack in the latter and this is visible when you see the interactions of their VP's. Luckily for them, if you start to create enough of a lead with the tech, you get many chances to sort out the product stuff.
      • dakolli 8 hours ago
        You sound like Russ Hanneman from SV
        • s-kymon 8 hours ago
          It's not about how much you earn. It's about what you're worth.
    • Razengan 8 hours ago
      Gemini's UX (and of course privacy cred as with anything Google) is the worst of all the AI apps. In the eyes of the Common Man, it's UI that will win out, and ChatGPT's is still the best.
      • xnx 7 hours ago
        Google privacy cred is ... excellent? The worst data breach I know of them having was a flaw that allowed access to names and emails of 500k users.
        • oofbey 4 minutes ago
          Their SECURITY cred is fantastic.

          Privacy, not so much. How many hundreds of millions have they been fined for “incognito mode” in chrome being a blatant lie?

        • bitpush 7 hours ago
          Link? Are you conflating with "500k Gmail accounts leaked [by a third party]" with Gmail having a breach?

          Afaik, Google has had no breaches ever.

        • laurex 7 hours ago
          If you consider "privacy" to be 'a giant corporation tracks every bit of possible information about you and everyone else'?
          • xnx 5 hours ago
            OpenAI is running ads. Do you think they'll track less?
        • Razengan 7 hours ago
          They don't even let you have multiple chats if you disable their "App Activity" or whatever (wtf is with that ass naming? they don't even have a "Privacy" section in their settings the last time I checked)

          and when I swap back into the Gemini app on my iPhone after a minute or so the chat disappears. and other weird passive-aggressive take-my-toys-away behavior if you don't bare your body and soul to Googlezebub.

          ChatGPT and Grok work so much better without accounts or with high privacy settings.

      • alexpotato 7 hours ago
        > Gemini's UX ... is the worst of all the AI apps

        Been using Gemini + OpenCode for the past couple weeks.

        Suddenly, I get a "you need a Gemini Access Code license" error but when you go to the project page there is no mention of this or how to get the license.

        You really feel the "We're the phone company and we don't care. Why? Because we don't have to." [0] when you use these Google products.

        PS for those that don't get the reference: US phone companies in the 1970s had a monopoly on local and long distance phone service. Similar to Google for search/ads (really a "near" monopoly but close enough).

        0 - https://vimeo.com/355556831

      • ainch 4 hours ago
        I find Gemini's web page much snappier to use than ChatGPT - I've largely swapped to it for most things except more agentic tasks.
      • jonathanstrange 8 hours ago
        You mean AI Studio or something like that, right? Because I can't see a problem with Google's standard chat interface. All other AI offerings are confusing both regarding their intended use and their UX, though, I have to concur with that.
        • ergonaught 7 hours ago
          The lack of "projects" alone makes their chat interface really unpleasant compared to ChatGPT and Claude.
        • xnx 7 hours ago
          AI Studio is also significantly improved as of yesterday.
        • wiseowise 7 hours ago
          No projects, completely forgets context mid dialog, mediocre responses even on thinking, research got kneecapped somehow and is completely uses now, uses propaganda Russian videos as the search material (what’s wrong with you, Google?), janky on mobile, consumes GIGABYTES of RAM on web (seriously, what the fuck?). Left a couple of tabs over night, Mac is almost complete frozen because 10 tabs consumed 8 GBs of RAM doing nothing. It’s a complete joke.
      • uxhoiuewfhhiu 5 hours ago
        Gemini is completely unusable in VS Code. It's rated 2/5 stars, pathetic: https://marketplace.visualstudio.com/items?itemName=Google.g...

        Requests regularly time out, the whole window freezes, it gets stuck in schizophrenic loops, edits cannot be reverted and more.

        It doesn't even come close to Claude or ChatGPT.

        • xnx 12 minutes ago
          Once Google launched Antigravity, I stopped using VS Code.
        • Razengan 28 minutes ago
          Smart idea to say anything against Google here from a throwaway account, I'm sitting in negative karma for that :')
  • rob-wagner 4 hours ago
    I’ve been using Gemini 3 Pro on a historical document archiving project for an old club. One of the guys had been working on scanning old handwritten minutes books written in German that were challenging to read (1885 through 1974). Anyways, I was getting decent results on a first pass with 50 page chunks but ended up doing 1 page at a time (accuracy probably 95%). For each page, I submit the page for a transcription pass followed by a translation of the returned transcription. About 2370 pages and sitting at about $50 in Gemini API billing. The output will need manual review, but the time savings is impressive.
    • dubeye 3 hours ago
      It sounds like a job where one pass might also be a viable option. Until you do the manual review you won't have a full sense of the time savings involved.
      • websap 21 minutes ago
        They could likely increase their budget slightly and run an LLM-based judge.
      • rob-wagner 2 hours ago
        Good idea. I’ll try modifying the prompt to transcribe, identify the language, and translate if not English, and then return a structured result. In my spot checks, most of the errors are in people’s names and if the handwriting trails into margins (especially into the fold of the binding). Even with the data still needing review, the translations from it has revealed a lot of interesting characters as well as this little anecdote from the minutes of the June 6, 1941 Annual Meeting:

        It had already rained at the beginning of the meeting. During the same, however, a heavy thunderstorm set in, whereby our electric light line was put out of operation. Wax candles with beer bottles as light holders provided the lighting. In the meantime the rain had fallen in a cloudburst-like manner, so that one needed help to get one's automobile going. In some streets the water stood so high that one could reach one's home only by detours. In this night 9.65 inches of rain had fallen.

        • rjcrystal 14 minutes ago
          One discovery I've made with gemini is that ocr accuracy is much higher when document is perfectly aligned at 0 degree. When we provided images with handwritten text to gemini which were horizontal (90 or 180 degree) it had lots of issues reading dates, names etc. Then we used paddle ocr image orientation model to find orientation and rotate the image it solved most of our issues with ocr.
  • sigmar 10 hours ago
    Here is the methodologies for all the benchmarks: https://storage.googleapis.com/deepmind-media/gemini/gemini_...

    The arc-agi-2 score (84.6%) is from the semi-private eval set. If gemini-3-deepthink gets above 85% on the private eval set, it will be considered "solved"

    >Submit a solution which scores 85% on the ARC-AGI-2 private evaluation set and win $700K. https://arcprize.org/guide#overview

    • gs17 9 hours ago
      Interestingly, the title of that PDF calls it "Gemini 3.1 Pro". Guess that's dropping soon.
      • sigmar 9 hours ago
        I looked at the file name but not the document title (specifically because I was wondering if this is 3.1). Good spot.

        edit: they just removed the reference to "3.1" from the pdf

        • josalhor 8 hours ago
          I think this is 3.1 (3.0 Pro with the RL improv of 3.0 Flash). But they probably decided to market it as Deep Think because why not charge more for it.
          • WarmWash 8 hours ago
            The Deep Think moniker is for parallel compute models though, not long CoT like pro models.

            It's possible though that deep think 3 is running 3.1 models under the hood.

      • staticman2 8 hours ago
        That's odd considering 3.0 is still labeled a "preview" release.
        • ainch 4 hours ago
          I think it'll be 3.1 by the time it's labelled GA - they said after 3.0 launch that they figured out new RL methods for Flash that the Pro model hasn't benefitted from.
      • WarmWash 8 hours ago
        The rumor was that 3.1 was today's drop
    • riku_iki 8 hours ago
      > If gemini-3-deepthink gets above 85% on the private eval set, it will be considered "solved"

      They never will do on private set, because it would mean its being leaked to google.

  • simianwords 9 hours ago
    OT but my intuition says that there’s a spectrum

    - non thinking models

    - thinking models

    - best of N models like deep think an gpt pro

    Each one is of a certain computational complexity. Simplifying a bit, I think they map to - linear, quadratic and n^3 respectively.

    I think there are certain class of problems that can’t be solved without thinking because it necessarily involves writing in a scratchpad. And same for best of N which involves exploring.

    Two open questions

    1) what’s the higher level here, is there a 4th option?

    2) can a sufficiently large non thinking model perform the same as a smaller thinking?

    • futureshock 6 hours ago
      I think step 4 is the agent swarm. Manager model gets the prompt and spins up a swarm of looping subagents, maybe assigns them different approaches or subtasks, then reviews results, refines the context files and redeploys the swarm on a loop till the problem is solved or your credit card is declined.
      • simianwords 6 hours ago
        i think this is the right answer

        edit: i don't know how this is meaningfully different from 3

    • NitpickLawyer 9 hours ago
      > best of N models like deep think an gpt pro

      Yeah, these are made possible largely by better use at high context lengths. You also need a step that gathers all the Ns and selects the best ideas / parts and compiles the final output. Goog have been SotA at useful long context for a while now (since 2.5 I'd say). Many others have come with "1M context", but their usefulness after 100k-200k is iffy.

      What's even more interesting than maj@n or best of n is pass@n. For a lot of applications youc an frame the question and search space such that pass@n is your success rate. Think security exploit finding. Or optimisation problems with quick checks (better algos, kernels, infra routing, etc). It doesn't matter how good your pass@1 or avg@n is, all you care is that you find more as you spend more time. Literally throwing money at the problem.

    • mnicky 9 hours ago
      > can a sufficiently large non thinking model perform the same as a smaller thinking?

      Models from Anthropic have always been excellent at this. See e.g. https://imgur.com/a/EwW9H6q (top-left Opus 4.6 is without thinking).

      • simianwords 9 hours ago
        its interesting that opus 4.6 added a paramter to make it think extra hard.
  • Scene_Cast2 7 hours ago
    It's a shame that it's not on OpenRouter. I hate platform lock-in, but the top-tier "deep think" models have been increasingly requiring the use of their own platform.
    • raybb 7 hours ago
      OpenRouter is pretty great but I think litellm does a very good job and it's not a platform middle man, just a python library. That being said, I have tried it with the deep think models.

      https://docs.litellm.ai/docs/

      • imiric 5 hours ago
        Part of OpenRouter's appeal to me is precisely that it is a middle man. I don't want to create accounts on every provider, and juggle all the API keys myself. I suppose this increases my exposure, but I trust all these providers and proxies the same (i.e. not at all), so I'm careful about the data I give them to begin with.
        • octoberfranklin 4 hours ago
          Unfortunately that's ending with mandatory-BYOK from the model vendors. They're starting to require that you BYOK to force you through their arbitrary+capricious onboarding process.
    • chr15m 3 hours ago
      The golden age is over.
  • jetter 7 hours ago
    it is interesting that the video demo is generating .stl model. I run a lot of tests of LLMs generating OpenSCAD code (as I have recently launched https://modelrift.com text-to-CAD AI editor) and Gemini 3 family LLMs are actually giving the best price-to-performance ratio now. But they are very, VERY far from being able to spit out a complex OpenSCAD model in one shot. So, I had to implement a full fledged "screenshot-vibe-coding" workflow where you draw arrows on 3d model snapshot to explain to LLM what is wrong with the geometry. Without human in the loop, all top tier LLMs hallucinate at debugging 3d geometry in agentic mode - and fail spectacularly.
    • mchusma 6 hours ago
      Hey, my 9 year old son uses modelrift for creating things for his 3d printer, its great! Product feedback: 1. You should probably ask me to pay now, I feel like i've used it enough. 2. You need a main dashboard page with a history of sessions. He thought he lost a file and I had to dig in the billing history to get a UUID I thought was it and generate the url. I would say naming sessions is important, and could be done with small LLM after the users initial prompt. 3. I don't think I like the default 3d model in there once I have done something, blank would be better.

      We download the stl and import to bambu. Works pretty well. A direct push would be nice, but not necessary.

    • gundmc 6 hours ago
      Yes, I've been waiting for a real breakthrough with regard to 3D parametric models and I don't think think this is it. The proprietary nature of the major players (Creo, Solidworks, NX, etc) is a major drag. Sure there's STP, but there's too much design intent and feature loss there. I don't think OpenSCAD has the critical mass of mindshare or training data at this point, but maybe it's the best chance to force a change.
    • lern_too_spel 6 hours ago
      If you want that to get better, you need to produce a 3d model benchmark and popularize it. You can start with a pelican riding a bicycle with working bicycle.
  • Decabytes 6 hours ago
    Gemini has always felt like someone who was book smart to me. It knows a lot of things. But if you ask it do anything that is offscript it completely falls apart
    • dwringer 5 hours ago
      I strongly suspect there's a major component of this type of experience being that people develop a way of talking to a particular LLM that's very efficient and works well for them with it, but is in many respects non-transferable to rival models. For instance, in my experience, OpenAI models are remarkably worse than Google models in basically any criterion I could imagine; however, I've spent most of my time using the Google ones and it's only during this time that the differences became apparent and, over time, much more pronounced. I would not be surprised at all to learn that people who chose to primarily use Anthropic or OpenAI models during that time had an exactly analogous experience that convinced them their model was the best.
    • piyh 58 minutes ago
      I made offmetaedh.com with it. Feels pretty great to me.
    • esafak 6 hours ago
      I'd rather say it has a mind of its own; it does things its way. But I have not tested this model, so they might have improved its instruction following.
      • vkazanov 6 hours ago
        Well, one thing i know for sure: it reliably misplaces parentheses in lisps.
        • esafak 5 hours ago
          Clearly, the AI is trying to steer you towards the ML family of languages for its better type system, performance, and concurrency ;)
  • anematode 4 hours ago
    It found a small but nice little optimization in Stockfish: https://github.com/official-stockfish/Stockfish/pull/6613

    Previous models including Claude Opus 4.6 have generally produced a lot of noise/things that the compiler already reliably optimizes out.

  • ggregoire 1 hour ago
    Do we know what model is used by Google Search to generate the AI summary?

    I've noticed this week the AI summary now has a loader "Thinking…" (no idea if it was already there a few weeks ago). And after "Thinking…" it says "Searching…" and shows a list of favicons of popular websites (I guess it's generating the list of links on the right side of the AI summary?).

  • Metacelsus 10 hours ago
    According to benchmarks in the announcement, healthily ahead of Claude 4.6. I guess they didn't test ChatGPT 5.3 though.

    Google has definitely been pulling ahead in AI over the last few months. I've been using Gemini and finding it's better than the other models (especially for biology where it doesn't refuse to answer harmless questions).

    • CuriouslyC 8 hours ago
      Google is way ahead in visual AI and world modelling. They're lagging hard in agentic AI and autonomous behavior.
    • throwup238 9 hours ago
      The general purpose ChatGpt 5.3 hasn’t been released yet, just 5.3-codex.
    • neilellis 9 hours ago
      It's ahead in raw power but not in function. Like it's got the worlds fast engine but one gear! Trouble is some benchmarks only measure horse power.
      • NitpickLawyer 9 hours ago
        > Trouble is some benchmarks only measure horse power.

        IMO it's the other way around. Benchmarks only measure applied horse power on a set plane, with no friction and your elephant is a point sphere. Goog's models have always punched over what benchmarks said, in real world use @ high context. They don't focus on "agentic this" or "specialised that", but the raw models, with good guidance are workhorses. I don't know any other models where you can throw lots of docs at it and get proper context following and data extraction from wherever it's at to where you'd need it.

    • scarmig 6 hours ago
      > especially for biology where it doesn't refuse to answer harmless questions

      Usually, when you decrease false positive rates, you increase false negative rates.

      Maybe this doesn't matter for models at their current capabilities, but if you believe that AGI is imminent, a bit of conservatism seems responsible.

    • nkzd 8 hours ago
      Google models and CLI harness feels behind in agentic coding compared OpenAI and Antrophic
    • Davidzheng 9 hours ago
      I gather that 4.6 strengths are in long context agentic workflows? At least over Gemini 3 pro preview, opus 4.6 seems to have a lot of advantages
      • verdverm 9 hours ago
        It's a giant game of leapfrog, shift or stretch time out a bit and they all look equivalent
    • simianwords 10 hours ago
      The comparison should be with GPT 5.2 pro which has been used successfully to solve open math problems.
  • aliljet 7 hours ago
    The problem here is that it looks like this is released with almost no real access. How are people using this without submitting to a $250/mo subscription?
    • andxor 6 hours ago
      People are paying for the subscriptions.
    • tootie 5 hours ago
      I gather this isn't intended a consumer product. It's for academia and research institutions.
  • siva7 8 hours ago
    I can't shake of the feeling that Googles Deep Think Models are not really different models but just the old ones being run with higher number of parallel subagents, something you can do by yourself with their base model and opencode.
    • Davidzheng 8 hours ago
      And after i do that, how do i combine the output of 1000 subagents into one output? (Im not being snarky here, i think it's a nontrivial problem)
      • tifik 7 hours ago
        The idea is that each subagent is focused on a specific part of the problem and can use its entire context window for a more focused subtask than the overall one. So ideally the results arent conflicting, they are complimentary. And you just have a system that merges them.. likely another agent.
      • mattlondon 8 hours ago
        You just pipe it to another agent to do the reduce step (i.e. fan-in) of the mapreduce (fan-out)

        It's agents all the way down.

      • jonathanstrange 7 hours ago
        Start with 1024 and use half the number of agents each turn to distill the final result.
  • czhu12 3 hours ago
    It’s incredible how fast these models are getting better. I thought for sure a wall would be hit, but these numbers smashes previous benchmarks. Anyone have any idea what the big unlock that people are finding now?
    • fsh 2 hours ago
      Companies are optimizing for all the big benchmarks. This is why there is so little correlation between benchmark performance and real world performance now.
      • czhu12 2 hours ago
        Isn’t there? I mean, Claude code has been my biggest usecase and it basically one shots everything now
        • fsh 1 hour ago
          Yes, LLMs have become extremely good at coding (not software engineer though). But try using them for anything original that cannot be adapted from GitHub and Stack Overflow. I haven't seen much improvement at all at such tasks.
  • simonw 8 hours ago
    The pelican riding a bicycle is excellent. I think it's the best I've seen.

    https://simonwillison.net/2026/Feb/12/gemini-3-deep-think/

    • tasuki 5 hours ago
      Tbh they'd have to be absolutely useless at benchmarkmaxxing if they didn't include your pelican riding a bicycle...
    • nickthegreek 6 hours ago
      I routinely check out the pelicans you post and I do agree, this is the best yet. It seemed to me that the wings/arms were such a big hangup for these generators.
    • steve_adams_86 2 hours ago
      We've reached PGI
    • Manabu-eo 8 hours ago
      How likely this problem is already on the training set by now?
      • throwup238 8 hours ago
        For every combination of animal and vehicle? Very unlikely.

        The beauty of this benchmark is that it takes all of two seconds to come up with your own unique one. A seahorse on a unicycle. A platypus flying a glider. A man’o’war piloting a Portuguese man of war. Whatever you want.

        • recursive 8 hours ago
          No, not every combination. The question is about the specific combination of a pelican on a bicycle. It might be easy to come up with another test, but we're looking at the results from a particular one here.
          • svara 8 hours ago
            More likely you would just train for emitting svg for some description of a scene and create training data from raster images.
            • recursive 3 hours ago
              None of this works if the testers are collaborating with the trainers. The tests ostensibly need to be arms-length from the training. If the trainers ever start over-fitting to the test, the tester would come up with some new test secretly.
      • simonw 8 hours ago
        If anyone trains a model on https://simonwillison.net/tags/pelican-riding-a-bicycle/ they're going to get some VERY weird looking pelicans.
        • suddenlybananas 7 hours ago
          Why would they train on that? Why not just hire someone to make a few examples.
          • simonw 7 hours ago
            I look forward to them trying. I'll know when the pelican riding a bicycle is good but the ocelot riding a skateboard sucks.
            • suddenlybananas 7 hours ago
              But they could just train on an assortment of animals and vehicles. It's the kind of relatively narrow domain where NNs could reasonably interpolate.
              • simonw 7 hours ago
                The idea that an AI lab would pay a small army of human artists to create training data for $animal on $transport just to cheat on my stupid benchmark delights me.
                • suddenlybananas 7 hours ago
                  When you're spending trillions on capex, paying a couple of people to make some doodles in SVGs would not be a big expense.
                  • simonw 6 hours ago
                    The embarrassment of getting caught doing that would be expensive.
      • zarzavat 8 hours ago
        You can always ask for a tyrannosaurus driving a tank.
      • verdverm 8 hours ago
        I've heard it posited that the reason the frontier companies are frontier is because they have custom data and evals. This is what I would do too
    • enraged_camel 7 hours ago
      Is there a list of these for each model, that you've catalogued somewhere?
    • throwup238 8 hours ago
      The reflection of the sun in the water is completely wrong. LLMs are still useless. (/s)
    • deron12 8 hours ago
      It's worth noting that you mean excellent in terms of prior AI output. I'm pretty sure this wouldn't be considered excellent from a "human made art" perspective. In other words, it's still got a ways to go!

      Edit: someone needs to explain why this comment is getting downvoted, because I don't understand. Did someone's ego get hurt, or what?

      • gs17 8 hours ago
        It depends, if you meant from a human coding an SVG "manually" the same way, I'd still say this is excellent (minus the reflection issue). If you meant a human using a proper vector editor, then yeah.
        • fvdessen 7 hours ago
          maybe you're a pro vector artist but I couldn't create such a cool one myself in illustrator tbh
      • dfdsf2 8 hours ago
        Indeed. And when you factor in the amount invested... yeah it looks less impressive. The question is how much more money needs to be invested to get this thing closer to reality? And not just in this instance. But for any instance e.g. a seahorse on a bike.
    • saberience 8 hours ago
      Do you have to still keep trying to bang on about this relentlessly?

      It was sort of humorous for the maybe first 2 iterations, now it's tacky, cheesy, and just relentless self-promotion.

      Again, like I said before, it's also a terrible benchmark.

      • jeanloolz 5 hours ago
        I'll agree to disagree. In any thread about a new model, I personally expect the pelican comment to be out there. It's informative, ritualistic and frankly fun. Your comment however, is a little harsh. Why mad?
      • simonw 7 hours ago
        It being a terrible benchmark is the bit.
      • Davidzheng 8 hours ago
        Eh, i find it more of a not very informative but lighthearted commentary
    • dfdsf2 8 hours ago
      Highly disagree.

      I was expecting something more realistic... the true test of what you are doing is how representative is the thing in relation to the real world. E.g. does the pelican look like a pelican as it exists in reality? This cartoon stuff is cute but doesnt pass muster in my view.

      If it doesn't relate to the real world, then it most likely will have no real effect on the real economy. Pure and simple.

      • chriswarbo 8 hours ago
        I disagree. The task asks for an SVG; which is a vector format associated with line drawings, clipart and cartoons. I think it's good that models are picking up on that context.

        In contrast, the only "realistic" SVGs I've seen are created using tools like potrace, and look terrible.

        I also think the prompt itself, of a pelican on bicycle, is unrealistic and cartoonish; so making a cartoon is a good way to solve the task.

      • peaseagee 8 hours ago
        The request is for an SVG, generally _not_ the format for photorealistic images. If you want to start your own benchmark, feel free to ask for a photorealistic JPEG or PNG of a pelican riding a bicycle. Could be interesting to compare and contrast, honestly.
  • mark_l_watson 3 hours ago
    I feel like a luddite: unless I am running small local models, I use gemini-3-flash for almost everything: great for tool use, embedded use in applications, and Python agentic libraries, broad knowledge, good built in web search tool, etc. Oh, and it is fast and cheap.

    I really only use gemini-3-pro occasionally when researching and trying to better understand something. I guess I am not a good customer for super scalers. That said, when I get home from travel, I will make a point of using Gemini 3 Deep Think for some practical research. I need a business card with the title "Old Luddite."

  • mark_l_watson 3 hours ago
    Off topic comment (sorry): when people bash "models that are not their favorite model" I often wonder if they have done the engineering work to properly use the other models. Different models and architectures often require very different engineering to properly use them. Also, I think it is fine and proper that different developers prefer different models. We are in early days and variety is great.
  • neilellis 9 hours ago
    Less than a year to destroy Arc-AGI-2 - wow.
    • Davidzheng 9 hours ago
      I unironically believe that arc-agi-3 will have a introduction to solved time of 1 month
      • ACCount37 6 hours ago
        Not very likely?

        ARC-AGI-3 has a nasty combo of spatial reasoning + explore/exploit. It's basically adversarial vs current AIs.

      • etyhhgfff 8 hours ago
        The AGI bar has to be set even higher, yet again.
      • dakolli 7 hours ago
        wow solving useless puzzles, such a useful metric!
        • esafak 5 hours ago
          How is spatial reasoning useless??
    • modeless 7 hours ago
      It's still useful as a benchmark of cost/efficiency.
    • XCSme 8 hours ago
      But why only a +0.5% increase for MMMU-Pro?
      • kingstnap 6 hours ago
        Its possibly label noise. But you can't tell from a single number.

        You would need to check to see if everyone is having mistakes on the same 20% or different 20%. If its the same 20% either those questions are really hard, or they are keyed incorrectly, or they aren't stated with enough context to actually solve the problem.

        It happens. Old MMLU non pro had a lot of wrong answers. Simple things like MNIST have digits labeled incorrect or drawn so badly its not even a digit anymore.

      • kenjackson 7 hours ago
        Everyone is already at 80% for that one. Crazy that we were just at 50% with GPT-4o not that long ago.
        • XCSme 2 hours ago
          But 80% sounds far from good enough, that's 20% error rate, unusable in autonomous tasks. Why stop at 80%? If we aim for AGI, it should 100% any benchmark we give.
          • kenjackson 2 hours ago
            Are humans 100%?
            • XCSme 2 hours ago
              If they are knowledgeable enough and pay attention, yes. Also, if they are given enough time for the task.

              But the idea of automation is to make a lot fewer mistakes than a human, not just to do things faster and worse.

    • saberience 8 hours ago
      It's a useless meaningless benchmark though, it just got a catchy name, as in, if the models solve this it means they have "AGI", which is clearly rubbish.

      Arc-AGI score isn't correlated with anything useful.

      • Legend2440 5 hours ago
        It's correlated with the ability to solve logic puzzles.

        It's also interesting because it's very very hard for base LLMs, even if you try to "cheat" by training on millions of ARC-like problems. Reasoning LLMs show genuine improvement on this type of problem.

      • jabedude 8 hours ago
        how would we actually objectively measure a model to see if it is AGI if not with benchmarks like arc-AGI?
        • WarmWash 7 hours ago
          Give it a prompt like

          >can u make the progm for helps that with what in need for shpping good cheap products that will display them on screen and have me let the best one to get so that i can quickly hav it at home

          And get back an automatic coupon code app like the user actually wanted.

      • HDThoreaun 3 hours ago
        ARC-AGI 2 is an IQ test. IQ tests have been shown over and over to have predictive power in humans. People who score well on them tend to be more successful
        • fsh 2 hours ago
          IQ tests only work if the participants haven't trained for them. If they do similar tests a few times in a row, scores increase a lot. Current LLMs are hyper-optimized for the particular types of puzzles contained in popular "benchmarks".
  • sinuhe69 8 hours ago
    I'm pretty certain that DeepMind (and all other labs) will try their frontier (and even private) models on First Proof [1].

    And I wonder how Gemini Deep Think will fare. My guess is that it will get half the way on some problems. But we will have to take an absence as a failure, because nobody wants to publish a negative result, even though it's so important for scientific research.

    [1] https://1stproof.org/

    • zozbot234 8 hours ago
      The 1st proof original solutions are due to be published in about 24h, AIUI.
    • octoberfranklin 4 hours ago
      Really surprised that 1stproof.org was submitted three times and never made front page at HN.

      https://hn.algolia.com/?q=1stproof

      This is exactly the kind of challenge I would want to judge AI systems based on. It required ten bleeding-edge-research mathematicians to publish a problem they've solved but hold back the answer. I appreciate the huge amount of social capital and coordination that must have taken.

      I'm really glad they did it.

  • ramshanker 9 hours ago
    Do we get any model architecture details like parameter size etc.? Few months back, we used to talk more on this, now it's mostly about model capabilities.
    • Davidzheng 9 hours ago
      I'm honestly not sure what you mean? The frontier labs have kept arch as secrets since gpt3.5
      • willis936 6 hours ago
        At the very least gemini 3's flyer claims 1T parameters.
  • ipaddr 3 hours ago
    The benchmark should be: can you ask it to create a profitable business or product and send you the profit?

    Everything else is bike shedding.

  • vessenes 9 hours ago
    Not trained for agentic workflows yet unfortunately - this looks like it will be fantastic when they have an agent friendly one. Super exciting.
    • dakolli 8 hours ago
      Its really weird how you all are begging to be replaced by llms, you think if agentic workflows get good enough you're going to keep your job? Or not have your salary reduced by 50%?

      If Agents get good enough it's not going to build some profitable startup for you (or whatever people think they're doing with the llm slot machines) because that implies that anyone else with access to that agent can just copy you, its what they're designed to do... launder IP/Copyright. Its weird to see people get excited for this technology.

      None of this good. We are simply going to have our workforces replaced by assets owned by Google, Anthropic and OpenAI. We'll all be fighting for the same barista jobs, or miserable factory jobs. Take note on how all these CEOs are trying to make it sound cool to "go to trade school" or how we need "strong American workers to work in factories".

      • BeetleB 5 hours ago
        > Its really weird how you all are begging to be replaced by llms, you think if agentic workflows get good enough you're going to keep your job? Or not have your salary reduced by 50%?

        The computer industry (including SW) has been in the business of replacing jobs for decades - since the 70's. It's only fitting that SW engineers finally become the target.

      • ergonaught 7 hours ago
        Most folks don't seem to think that far down the line, or they haven't caught on to the reality that the people who actually make decisions will make the obvious kind of decisions (ex: fire the humans, cut the pay, etc) that they already make.
        • blibble 4 hours ago
          they think they're going to be the person making that decision

          but forgot there's likely someone above them making exactly the same one about them

      • sgillen 6 hours ago
        I think a lot of people assume they will become highly paid Agent orchestrators or some such. I don't think anyone really knows where things are heading.
        • dakolli 1 hour ago
          Why would someone get paid well for this skill? Its not valuable at all.
      • timeattack 5 hours ago
        I agree with you and have similar thoughts (maybe, unfortunately for me). I personally know people who outsource not just their work, but also their life to LLMs, and reading their exciting comments makes me feel a mix of cringe, fomo and dread. But what is the engame for me and you likes, when we finally would be evicted from our own craft? Stash money while we still can, watching 'world crash and burn', and then go and try to ascend in some other, not yet automated craft?
        • dakolli 4 hours ago
          Yeah, that's a good question that I can't stop thinking about. I don't really enjoy much else other than building software, its genuinely my favorite thing to do. Maybe there will be a world where we aren't completely replaced, we have handmade clothes still after all that are highly coveted. I just worry its going to uproot more than just software engineering, theoretically it shouldn't be hard to replace all low hanging fruit in the realm of anything that deals with computer I/O. Previous generations of automation have created new opportunities for humans, but this seems mostly just as a means of replacement. The advent of mass transportation/vehicles created machines who needed mechanics (and eventually software), I don't see that happening in this new paradigm.

          I don't think that's going to make society very pleasant if everyone's fighting over the few remaining ways to make livelihood. People need to work to eat. I certainly don't see the capitalist class giving everyone UBI and letting us garden or paint for the rest of our lives. I worry we're likely going to end up in trenches or purged through some other means.

      • vessenes 4 hours ago
        I’m someone who’d like to deploy a lot more workers than I want to manage.

        Put another way, I’m on the capital side of the conversation.

        The good news for labor that has experience and creativity is that it just started costing 1/100,000 what it used to to get on that side of the equation.

        • jimmymcgee73 2 hours ago
          If LLMs truly cause widespread replacement of labor, you’re screwed just as much as anyone else. If we hit say 40% unemployment do you think people will care you own your home or not? Do you think people will care you have currency or not? The best case outcome will be universal income and a pseudo utopia where everyone does ok. The “bad” scenario is widespread war.

          I am one of the “haves” and am not looking forward to the instability this may bring. Literally no one should.

          • blibble 2 hours ago
            > I am one of the “haves” and am not looking forward to the instability this may bring. Literally no one should.

            these people always forget capitalism is permitted to exist by consent of the people

            if there's 40% unemployment it won't continue to exist, regardless of what the TV/tiktok/chatgpt says

          • dakolli 1 hour ago
            Well he also thinks $10.00 in LLM tokens is equivalent to a $1mm labor budget. These are the same people who were grifting during the NFTs days, claiming they were the future of art.
        • dakolli 1 hour ago
          lmao, you are an idealistic moron. If llms can replace labor at 1/100k of the cost (lmfao) why are you looking to "deploy" more workers? So are you trying to say if I have $100.00 in tokens I have the equivalent of $10mm in labor potential.... What kind of statement is this?

          This is truly the dumbest statement I've ever seen on this site for too many reasons to list.

          You people sound like NFT people in 2021 telling people that they're creating and redefining art.

          Oh look peter@capital6.com is a "web3" guy. Its all the same grifters from the NFT days behaving the same way.

      • newswasboring 6 hours ago
        You don't hate AI, you hate capitalism. All the problems you have listed are not AI issues, its this crappy system where efficiency gains always end up with the capital owners.
      • OtomotO 8 hours ago
        [flagged]
        • dakolli 8 hours ago
          Well I honestly think this is the solution. It's much harder to do French Revolution V2 though if they've used ML to perfect people's recommendation algorithms to psyop them into fighting wars on behalf of capitalists.

          I imagine llm job automation will make people so poor that they beg to fight in wars, and instead of turning that energy against he people who created the problem they'll be met with hours of psyops that direct that energy to Chinese people or whatever.

          We will see.

        • uxhoiuewfhhiu 5 hours ago
          [flagged]
  • sega_sai 4 hours ago
    I do like google models (and I pay for them), but the lack of competitive agent is a major flaw in Google's offering. It is simply not good enough in comparison to claude code. I wish they put some effort there (as I don't want to pay two subscriptions to both google and anthropic)
  • dmbche 2 hours ago
    So what happens if the AI companies can't make money? I see more and more advances and breakthrough but they are taking in debt and no revenue in sight.

    I seem to understand debt is very bad here since they could just sell more shares, but aren't (either valuation is stretched or no buyers).

    Just a recession? Something else? Aren't they very very big to fall?

    Edit0: Revenue isn't the right word, profit is more correct. Amazon not being profitable fucks with my understanding of buisness. Not an economist.

    • sigmar 2 hours ago
      >taking in debt and no revenue in sight.

      which companies don't have revenue? anthropic is at a run rate of 14 billion (up from 9B in December, which was up from 4B in July). Did you mean profit? They expect to be cash flow positive in 2028.

      • dmbche 28 minutes ago
        Yes thank you, mixing my brushes here - I remembered one of the companies having raised over 100b and having about 10b in revenue.
    • echelon 2 hours ago
      AI will kill SaaS moats and thus revenue. Anyone can build new SaaS quickly. Lots of competition will lead to marginal profits.

      AI will kill advertising. Whatever sits at the top "pane of glass" will be able to filter ads out. Personal agents and bots will filter ads out.

      AI will kill social media. The internet will fill with spam.

      AI models will become commodity. Unless singularity, no frontier model will stay in the lead. There's competition from all angles. They're easy to build, just capital intensive (though this is only because of speed).

      All this leaves is infrastructure.

      • ddxv 1 hour ago
        Not following some of the jumps here.

        Advertising, how will they kill ads any better than the current cat and mouse games with ad blockers?

        Social Media, how will they kill social media? Probably 80% of the LinkedIn posts are touched by AI (lots of people spend time crafting them, so even if AI doesn't write the whole thing you know they ran the long ones through one) but I'm still reading (ok maybe skimming) the posts.

        • echelon 1 hour ago
          > Advertising, how will they kill ads any better than the current cat and mouse games with ad blockers?

          The Ad Blocker cat and mouse game relies on human-written metaheuristics and rules. It's annoying for humans to keep up. It's difficult to install.

          Agents/Bots or super slim detection models will easily be able to train on ads and nuke them whatever form they come in: javascript, inline DOM, text content, video content.

          Train an anti-Ad model and it will cleanse the web of ads. You just need a place to run it from the top.

          You wouldn't even have to embed this into a browser. It could run in memory with permissions to overwrite the memory of other applications.

          > Social Media, how will they kill social media?

          MoltClawd was only the beginning. Soon the signal will become so noisy it will be intolerable. Just this week, X's Nikita Bier suggested we have less than six months before he sees no solution.

          Speaking of X, they just took down Higgsfield's (valued at $1.3B) main account because they were doing it across a molt bot army, and they're not the only ones. Extreme measures were the only thing they could do. For the distributed spam army, there will be no fix. People are already getting phone calls from this stuff.

    • ipnon 2 hours ago
      They're using the ride share app playbook. Subsidize the product to reach market saturation. Once you've found a market segment that depends on your product you raise the price to break even. One major difference though is that ride share's haven't really changed in capabilities since they launched: it's a map that shows a little car with your driver coming and a pin where you're going. But it's reasonable to believe that AI will have new fundamental capabilities in the 2030s, 2040s, and so on.
  • nphardon 2 hours ago
    I think I'm finally realizing that my job probably won't exist in 3-5. Things are moving so fast now that the LLMs are basically writing themselves. I think the earlier iterations moved slower because they were limited by human ability and productivity limitations.
  • Legend2440 5 hours ago
    I'm really interested in the 3D STL-from-photo process they demo in the video.

    Not interested enough to pay $250 to try it out though.

  • Dirak 6 hours ago
    Praying this isn't another Llama4 situation where the benchmark numbers are cooked. 84.6% on Arc-AGI is incredible!
  • ismailmaj 8 hours ago
    top 10 elo in codeforces is pretty absurd
  • jonathanstrange 9 hours ago
    Unfortunately, it's only available in the Ultra subscription if it's available at all.
  • andrewstuart 8 hours ago
    Gemini was awesome and now it’s garbage.

    It’s impossible for it to do anything but cut code down, drop features, lose stuff and give you less than the code you put in.

    It’s puzzling because it spent months at the head of the pack now I don’t use it at all because why do I want any of those things when I’m doing development.

    I’m a paid subscriber but there’s no point any more I’ll spend the money on Claude 4.6 instead.

    • halapro 7 hours ago
      I never found it useful for code. It produced garbage littered with gigantic comments.

      Me: Remove comments

      Literally Gemini: // Comments were removed

      • andrewstuart 7 hours ago
        It would make more sense to me if it had never been awesome.
        • mortsnort 6 hours ago
          They may quantize the models after release to save money.
    • ergonaught 7 hours ago
      It seems to be adept at reviewing/editing/critiquing, at least for my use cases. It always has something valuable to contribute from that perspective, but has been comparatively useless otherwise (outside of moats like "exclusive access to things involving YouTube").
  • m3kw9 8 hours ago
    Gemini 3 Pro/Flash is stuck in preview for months now. Google is slow but they progress like a massive rock giant.
  • okokwhatever 8 hours ago
    I need to test the sketch creation a s a p. I need this in my life because learning to use Freecad is too difficult for a busy person like me (and frankly, also quite lazy)
    • sho_hn 8 hours ago
      FWIW, the FreeCAD 1.1 nightlies are much easier and more intuitive to use due to the addition of many on-canvas gizmos.
  • syntaxing 9 hours ago
    Why a Twitter post and not the official Google blog post… https://blog.google/innovation-and-ai/models-and-research/ge...
    • dang 8 hours ago
      Just normal randomness I suppose. I've put that URL at the top now, and included the submitted URL in the top text.
    • meetpateltech 9 hours ago
      The official blog post was submitted earlier (https://news.ycombinator.com/item?id=46990637), but somehow this story ranked up quickly on the homepage.
      • verdverm 9 hours ago
        @dang will often replace the post url & merge comments

        HN guidelines prefer the original source over social posts linking to it.

    • aavci 9 hours ago
      Agreed - blog post is more appropriate than a twitter post
  • kittbuilds 2 hours ago
    [dead]
  • bschmidt720 7 hours ago
    [dead]
  • HardCodedBias 7 hours ago
    Always the same with Google.

    Gemini has been way behind from the start.

    They use the firehose of money from search to make it as close to free as possible so that they have some adoption numbers.

    They use the firehose from search to pay for tons of researchers to hand hold academics so that their non-economic models and non-economic test-time-compute can solve isolated problems.

    It's all so tiresome.

    Try making models that are actually competitive, Google.

    Sell them on the actual market and win on actual work product in millions of people lives.

  • i_love_retros 1 hour ago
    Fuck google. Boycott Google.
    • rexpop 1 hour ago
      Israel is not one of the boots. Deplorable as their domestic policy may be, they're not wagging the dog of capitalist imperialism. To imply otherwise is to reveal yourself as biased, warped in a way that keeps you from going after much bigger, and more real systems of political economy holding back our civilization from universal human dignity and opportunity.
      • i_love_retros 37 minutes ago
        Lol what? Not sure if you are defending Israel or google because your communication style is awful. But if you are defending Israel then you're an idiot who is excusing genocide. If you're defending google then you're just a corporate bootlicker who means nothing.
  • dperhar 8 hours ago
    Does anyone actually use Gemini 3 now? I cant stand its sleek salesy way of introduction, and it doesnt hold to instructions hard – makes it unapplicable for MECE breakdowns or for writing.
    • sigmar 2 hours ago
      I use it often. Occasionally for quick questions, but mostly for deep research.
    • copperx 8 hours ago
      I do. It's excellent when paired with an MCP like context7.
    • throwa356262 8 hours ago
      I dont agree, Gemini 3 is pretty good, even the Lite version.
      • dperhar 8 hours ago
        What do you use it for and why? Genuinely curious
    • jeffbee 7 hours ago
      It indeed departs from instructions pretty regularly. But I find it very useful and for the price it beats the world.

      "The price" is the marginal price I am paying on top of my existing Google 1, YouTube Premium, and Google Fi subs, so basically nothing on the margin.

    • wetpaws 8 hours ago
      [dead]