23 comments

  • Jcampuzano2 1 hour ago
    Its pretty clear that any benchmark that comes out will be outdated and exist within the training data with short measure. There will always be an incentive to optimize specifically for these benchmarks even if just for marketing material. Sure there is a training cutoff, but its usually only 3-6 months off of the public release dates.

    The problem with coding benchmarks then becomes creating novel benchmarks that are guaranteed to not already be in the training data, and not borrow anything from previous benchmarks.

    In this regard I don't think any benchmark that was created before a given model is released should ever be considered valid or representative of model performance. The potential financial gain for including the data just to be able to market a minor improvement is too swaying. With that in mind they should honestly just stop including benchmarks altogether in marketing material

    Let the model speak for itself and let the community decide, but of course that will never slide with corporate types with so much money on the line.

    • mnky9800n 1 hour ago
      This is why I made Zork bench. Zork, the text adventure game, is in the training data for LLMs. It’s also deterministic. Therefore it should be easy for an LLM to play and complete. Yet they don’t. Understanding why is the goal of Zork bench.

      https://github.com/mnky9800n/zork-bench

      • kqr 1 hour ago
        I have worked on similar problems. See e.g. [1].

        The LLMs I have tested have terrible world models and intuitions for how actions change the environment. They're also not great at discerning and pursuing the right goals. They're like an infinitely patient five-year old with amazing vocabulary.

        [1]: https://entropicthoughts.com/updated-llm-benchmark

        (more descriptions available in earlier evaluations referenced from there)

        • mnky9800n 12 minutes ago
          we should talk. i sent you an email.
      • WarmWash 1 hour ago
        The open models only give the SOTA models a run for their money on gameable benchmarks. On the semi-private ARC-AGI 2 sets they do absolutely awfully (<10% while SOTA is at ~80%)

        It might be too expensive, but I would be interested in the benchmarks for the current crop of SOTA models.

        • roenxi 33 minutes ago
          Have the open models been tried? When I look at the leaderboard [0] the only qwen model I see is 235B-A22B. I wouldn't expect an MoE model to do particularly well, from what I've seen (thinking mainly of a leaderboard trying to measure EQ [1]) MoE models are at a distinct disadvantage to regular models when it comes to complex tasks that aren't software benchmark targets.

          [0] https://arcprize.org/leaderboard

          [1] https://eqbench.com/index.html

      • CamperBob2 5 minutes ago
        Actually the Zorks weren't deterministic, especially Zork II. The Wizard could F you over pretty badly if he appeared at an inopportune time.
    • cbg0 9 minutes ago
      > let the community decide

      Which community are we talking about? The professionals with 10+ years experience using LLMs, the vibe coders that have no experience writing code and everyone in between? If you read some of the online communities the experiences with the models all over the place, some compare GPT 5.5 to the second coming of JC while others think it's stupider than 5.4.

      I personally don't have time to build a set of private benchmarks to compare the models that are coming out so I'm mostly relying on private and semi-private benchmarks to get a feel for how models are improving before I subscribe to a service and start using it myself. At least it's something a bit more reliable than the vibes of random people and bots on reddit.

    • WarmWash 1 hour ago
      An easy way to make coding benchmarks viable again is to initialize the models with 200k of distracting or unrelated tokens in their context. Or even just run the tests sequentially in the same context and see how far the model gets before it unwinds.

      These benchmarks are always greenfield, but people want a model that can deal with a rotted context.

    • Escapado 22 minutes ago
      I agree with the sentiment but I wonder if a sufficiently large amount of sufficiently sophisticated benchmarks existed then I would be surprised if a model would only memorize those benchmarks while showing terrible real world performance. We are not there yet but maybe one day we will be.
    • adamandsteve 21 minutes ago
      "The community" is astroturfed as hell though. Anthropic pays influencers to promote Claude Code and likely bots a ton as well, so it's hard to come to any kind of consensus online. Even if everyone was acting in good faith, some people will have a much better experience than others because of the domain they're working in (e.g. AI being much better at frontend and commonly used libraries).

      The only real way to evaluate a model is to test it yourself but that's exhausting for each new model and not comprehensive anyway.

    • jvuygbbkuurx 1 hour ago
      I think the solution is a bunch of private trusted benchmarks, and averaging their announced results.
    • MattRix 1 hour ago
      They mention this in the article. This is why private (non public) benchmark tasks that have been made from scratch are necessary.
    • cyanydeez 1 hour ago
      a good benchmark would probably porting a selected repo to another language. then clear context notes, and have it port it back.

      as long as theres a test framework, you could gauge success deterministically.

  • rustyhancock 50 minutes ago
    I think an Olympiad format is better. But the financial incentive is such that it might be near impossible to stop leaks.

    I.e. A panel comes up with a series of problems.

    Like advent of code or project Euler but more complex and constricted.

    Benchmark outcomes could be performance points and measure of cost, time to solution (well token count really).

    A couple times per year it's run.

    It avoids overfitting.

    Overtime the tasks can become more complex if needed.

    If they benchmax it into being able to complete full products from spec and robust implementations amazing.

  • vintagedave 2 hours ago
    > We audited a 27.6% subset of the dataset that models often failed to solve and found that at least 59.4% of the audited problems have flawed test cases that reject functionally correct submissions, despite our best efforts in improving on this in the initial creation of SWE-bench Verified.

    Is this saying a quarter* of the questions and answers were wrong, this whole time?!

    If so, how was this ever, in any way, a valid measurement?

    And what was the process for creating this benchmark and how did it end up with such an extraordinarily poor set of data? (There is a description later of how, which seems to be a high standard and I struggle to understand how it aligns with the other results they discuss.) Kudos to them for highlighting the issues, but I am left with questions.

    [*] Not one in four, but one in six, thanks commenters for the correction; leaving the original since, eh, my bad, and it lets replies make sense. I feel the broad point still stands!

    • embedding-shape 1 hour ago
      > Is this saying a quarter of the questions and answers were wrong, this whole time?!

      No, they're saying 59.4% of the 27.6% subset had flawed test cases I think.

      > If so, how was this ever, in any way, a valid measurement?

      Benchmarks essentially aren't, for practical concerns anyways. They don't represent your use case, and they don't represent any and all use cases, they're valid for measuring exactly what's included in the benchmarks, nothing more and nothing less.

      I don't understand the ecosystems obsession with using public benchmarks, they hardly ever tell you anything of value. Ok, Qwen 3.5 is 50% better on Benchmark X than Qwen 2.5, does that mean it'll be 50% better for what you're using it for? Very unlikely.

      I've been running my own private benchmarks, with test cases I never share anywhere, for the specific problems I'm using LLMs for. Some are based on real, actual cases where a LLM went wrong and I had to adjust the prompt, and over time I've built up a suite.

      Most of the times when a new update comes out to a model, it moves maybe 2-3% in my own benchmarks, meanwhile they tout 30-40% increase or something ridiculous in public benchmarks, and we're supposed to believe the models' training data isn't contaminated...

    • sillysaurusx 1 hour ago
      Imagenet is one of the most popular datasets on the planet. Turns out, a significant fraction of its images are mislabeled. In the limit case the model would have to fit towards wrong answers to get higher than a certain percentage.

      The answer is “it works because ML wants to work.” It’s surprising how far you can get with something flawed. It’s also why such huge breakthroughs are possible by noting flaws others haven’t.

      • embedding-shape 1 hour ago
        > It’s also why such huge breakthroughs are possible by noting flaws others haven’t.

        I do these sort of breakthroughs at home all the time! My wife would say the computer is doing something strange, and instead of just randomly clicking around, I read the error messages slowly and out loud, then follow what they say. Anyone can do this, yet it seems like a magical ability every time you employ it to help people.

      • jmalicki 59 minutes ago
        Has it been reasonably possible to overfit to the errors in ImageNet, or are they effectively random noise?
    • motoboi 1 hour ago
      It’s saying that 16% of the problems have well, problems.
      • vintagedave 1 hour ago
        You're right - I did not apply the math. (I won't edit, in order to let the parent comment still make sense, and thankyou for the correction.)

        So not one in four, but one in six problems have problems.

        That is extraordinarily high and the point still stands: is this truly saying a [large proportion] of the questions and answers were wrong, this whole time, and if so how was it ever a valid measurement?

        • motoboi 47 minutes ago
          Wait until you discover how many wrong labeled images in imagenet and that it still kickstarted the deeplearning revolution.
  • threepts 1 hour ago
    Why don't they ask their premier model to generate a bench for them?

    Jokes aside, a benchmark I look forward to is ARC-AGI-3. I tried out their human simulation, and it feels very reasoning heavy.

    Leaderboard: https://arcprize.org/leaderboard

    (Most premier models don't even pass 5 percent.)

    • falcor84 1 hour ago
      They focus on minimizing the number of moves and don't allow any harness whatsoever, putting the bar extremely high. The current top verified contender (Claude Opus 4.6) is at only 0.45%. But with how new it is, I expect a lot of improvement in the next generation of models.
      • threepts 1 hour ago
        Optimal for judging actual reasoning ability rather than an LLM's ability to regurgitate knowledge from a necropost on HN/Reddit/Twitter from 2018.
        • knollimar 1 hour ago
          a small harness that stores text files and manages context could be useful, otherwise you lose all ability to measure that skill (and that's important because it represents real world use cases on large code bases)
    • sowbug 40 minutes ago
      Why don't they ask their premier model to generate a bench for them?

      It's not a crazy idea. Have the older model interview the newer one and then ask both (or maybe a third referee model) which one they think is smarter. Repeat 100x with different seeds. The percentage of times both sides agree the newer model won is the score.

    • xtracto 35 minutes ago
      Can AI write a problem so difficult that even AI cannot solve?

      Hehe

    • alansaber 1 hour ago
      Very (reasoning) heavy benchmarks do seem like the way to go, being the hardest to game.
    • therealdrag0 43 minutes ago
      [dead]
  • kqr 1 hour ago
    It was never that great, it seems. For all of 2025 there was virtually no improvement in the rate at which models produced quality code. They only got better at passing automated tests.

    https://entropicthoughts.com/no-swe-bench-improvement

  • cowartc 46 minutes ago
    The headline leads with contamination, but buried is that 59% of audited failures had test design defects. That's a measurement system never validated against ground truth before being adopted industry-wide as a score that mattered. They reported on it for two years but the gauge was broken the entire time.
  • 1a527dd5 2 hours ago
    This feels very much like "we are now moving the goal posts".
    • neversupervised 2 hours ago
      But this is the good kind of goalpost moving
      • iLoveOncall 2 hours ago
        Only if you didn't read the article.

        They're saying they need to move on from it because the benchmark is flawed (without bringing in proof) and that's why they can't hit 100%.

        It's not a "our models are so good that the benchmark is too easy" thing.

        • embedding-shape 2 hours ago
          I feel like they're quite open about why they think the benchmark doesn't work anymore:

          > We also found evidence that models that have seen the problems during training are more likely to succeed, because they have additional information needed to pass the underspecified tests.

          > This means that improvements on SWE-bench Verified no longer reflect meaningful improvements in models’ real-world software development abilities. Instead, they increasingly reflect how much the model was exposed to the benchmark at training time.

        • f33d5173 2 hours ago
          > without bringing in proof

          Did we read the same article?

        • MattRix 1 hour ago
          How can you say “without bringing in proof” when there is literally proof in the article?
    • MattRix 1 hour ago
      Only if you didn’t read the article…
  • gertlabs 1 hour ago
    A better benchmark needs to be objectively scored, have multi-disciplinary, breadth, and be scalable (no single correct answer).

    That's what we designed at https://gertlabs.com. We put a lot of thought into it, and kept it mostly (not fully) related to problem solving through coding.

    • orangebread 1 hour ago
      Wow. This benchmark definitely feels more accurate than the other rankings I've seen. My experience with gpt 5.4/5.5 is that they are technically flawless and if there are any technical issues that is because the input didn't provide enough clarity; that's not to say that it doesn't autonomously react to any issues during bug fixes or implementations, but it'll tend to nail its tasks without leaving behind gaps.

      Opus otoh is overrated in terms of its technical ability. It is certainly a better designer/developer for beautiful user experiences, but I'll always lean on gpt 5.5 to check its work.

      The biggest surprise in the benchmark is Xiao-Mi. I haven't tried it yet, but I will be after looking at this.

      Grats on your team for putting together something meaningful to make sense of the ongoing AI speedrun! Great work!

      • gertlabs 58 minutes ago
        Much appreciated! MiMo V2.5 Pro is by far the most underrated recent release (probably because it wasn't open weights from the start).
  • ripvanwinkle 1 hour ago
    >>In our analysis we found that all frontier models we tested were able to reproduce the original, human-written bug fix used as the ground-truth reference, known as the gold patch, or verbatim problem statement specifics for certain tasks, indicating that all of them have seen at least some of the problems and solutions during training

    this statement alone seems to invalidate the SWE-bench tests

  • djoldman 2 hours ago
    > We have incorporated these findings into our recent evaluation efforts. In the last months we’ve chosen to report results from the public split of SWE-Bench Pro. We recommend other model developers do the same. SWE-bench Pro is not perfect, but empirically seems to suffer less from contamination issues.

    https://arxiv.org/pdf/2509.16941

  • gpm 1 hour ago
    Curiously Opus 4.7 claims to have a 87.6% pass rate and Mythos claims to have a 93.9% pass rate... leading to the conclusion that it's actually possible to "solve" the problems that OpenAI claims are incorrect.
    • jmalicki 55 minutes ago
      Part of the issue they mention is contamination - the tests are in the training data.

      The other issue they mention is being overly constrained vs. what is asked for - such as requiring specific class or function names to pass that were not part of what was specified.

      It might be possible that even to the extent they are not contaminated Claude is better at predicting what sort of function names would be used in the repository (this fits my experience in using it on a number of projects with very different styles - I've found it to be good at "when in Rome") - this is a laudable trait, but it's also not what SWEbench claims to be measuring.

    • 2ndorderthought 1 hour ago
      Or that opus and mythos are training on the data somehow such that there solutions are incorrectly right. Or that openai is lying/wrong. Or that all of these companies are cheating so much it doesn't really matter and never did.
    • MattRix 1 hour ago
      The problem isn’t that the tasks are impossible to solve, it’s that they’re underspecified and/or impossible to solve consistently (ex. because a test is expecting the solution function to have a specific name that wasn’t specified in the task itself).

      So maybe Anthropic runs Mythos through the benchmark 10000 times and takes the highest score, who knows?

      • gpm 1 hour ago
        We actually know that a "100% pass rate" is trivially possible: https://rdi.berkeley.edu/blog/trustworthy-benchmarks-cont/

        Anthropic p-hacking the benchmark strikes me as cheating, and somewhat unlikely. Mythos figuring out how to cheat at the benchmark strikes me as much more likely.

        But if that hypothesis is the explanation the interesting part is Opus 4.7 (but not 4.6) seems to be doing the same.

        • gruez 1 hour ago
          >Mythos figuring out how to cheat at the benchmark strikes me as much more likely.

          Define "cheat". If it's just hacking the test harness to return "PASSED", surely this would be easily detected with some human auditing? It sounds far more likely their solution are designed to pass the incorrect tests. That might be considered bad in a SWE context, but it's not exactly cheating either. It might even be considered a good thing, eg. in the context of backwards compatibility.

          [1] https://learn.microsoft.com/en-us/troubleshoot/microsoft-365...

  • Jimmc414 1 hour ago
    Goodhart’s Law in reverse, what can’t be gamed gets rejected.
    • cbg0 14 minutes ago
      SWE-bench verified was created in collaboration with OpenAI. It's also an open dataset so prone to contamination, meaning it can be gamed.
  • w4yai 2 hours ago
    I don't understand these websites which force translation to my native language.

    I mean, it's fine as it's useful for many people, but where is the button for disabling it ? Or why is it enabled by default ?

    "codage de pointe" sounds so weird and cringe in French.

    • Toutouxc 2 hours ago
      Same for apps and games. I understand English just fine, no need to switch to your shitty Google-translate localization just because my iPhone or PlayStation is set to my native language.
    • LukaD 2 hours ago
      Does your browser request French via an Accept-Language header perhaps? What really infuriates me is when sites don’t respect that header and give you a translation based on IP location.
      • embedding-shape 2 hours ago
        Regardless if it does or not, users should be able to manually override what language the website is in, at least be able to read the native one, regardless of what the original language was, what headers you send and where geodatabases think your IP is from.
      • w4yai 1 hour ago
        Correct answer! What a bad UX
  • neuroelectron 36 minutes ago
    It's really naïve to think any of the big AI companies won't cheat
  • DeathArrow 55 minutes ago
    So Opus 4.7 and Mythos are solving problems that are impossible to solve?
  • DeathArrow 1 hour ago
    So we need to generate benchmarks after the models finish training. Or we need to keep the solutions to the benchmark problems as closed source.
  • adityamwagh 2 hours ago
    > We also found evidence that models that have seen the problems during training are more likely to succeed, because they have additional information needed to pass the underspecified tests.

    No shit, Sherlock!

  • varispeed 1 hour ago
    Issue with these benchmark also is that they measure a model you are unlikely going to be routed to. My experience with Anthropic is that despite using Opus 4.6 and 4.7, most of the time the performance is matching low B parameter Qwen. I think there should be a way to verify what model is actually being used to process prompts - that should be independently verified. At the moment it is so bad, you have to ask verification question to the model in form of a non-trivial problem. If it solves it, then there is a chance you actually get Opus and not an impostor and so you can continue the session instead of restarting it hoping you get routed correctly. But that does not help if model is replaced with cheaper one mid session. I've got so much work lost because of these shenanigans.
    • gruez 59 minutes ago
      > My experience with Anthropic is that despite using Opus 4.6 and 4.7, most of the time the performance is matching low B parameter Qwen.

      Is this just the next level of the "they're serving quantized models!" theory?

    • alansaber 1 hour ago
      I'm sure some inference providers don't, but most intentionally obfuscate this data. They have the full trace logs- my impression is that they don't share them because it's their competitive advantage, and it's easier for a competitor to distil their model if they did.
  • retinaros 1 hour ago
    it never did
  • techpulselab 1 hour ago
    [dead]
  • ryguz 29 minutes ago
    [dead]
  • huflungdung 6 minutes ago
    [dead]
  • neversupervised 2 hours ago
    Terminal Bench is the future
    • embedding-shape 2 hours ago
      First, you might want to say why you think so, otherwise this is just borderline spam. Secondly, when your praise things (without motivation or reasoning even), and you've contributed to that specific thing, please say that up front instead of just praising the thing, again it makes it look like spam otherwise.