At augment code we specifically build our code review tool to find noise to signal ratio problem. In benchmark our comments are 2 to 3x more likely to get fixed compared to bugbot coderabbit etc
At augmentcode.com, we've been evaluating Haiku for some time, it's actually a very good model. We found out it's 90% as good as Sonnet and is ~34% faster than sonnet!
Where it doesn't shine much is on very large coding task. but it is a phenomenal model for small coding tasks and the speed improvement is much welcome
90% as good as Sonnet 4 or 4.5?
Openrouter just started reporting, and it's saying Haiku is 2x as fast (60tps vs 125tps) and 2-3x less latent (2-3s vs 1s)
Fivetran acquired Census (reverse-etl) & Tobiko (dbt alternative).
I wonder who's next to really consolidate their platform play and compete with the old legacy MDM provider like Informatica. Data Observability or Catalog like Monte Carlo and Atlan. The whole Modern Data Stack has either died, acquired or merged by now. Wonder what's missing for Fivetran to IPO too.
I also wonder what this merge means for Airbyte who raised 150m at 1.5b in 2023.
Observability is a good guess, but I'd venture to guess that the conversations going on internally are about how to capture value across the entire stack. I wouldn't be surprised if we hear about them acquiring either a database/warehouse company and/or an analytics solution. Or vice versa, them getting acquired by a bigger player that wants to offer more connectors and data modeling functionality.
the MTR study is a joke. it surveyed only 16 devs. in the era of Sonnet 3.5
Can we stop citing this study
I'm not saying the DORA study is more accurate, but at least it surveyed 5000 developers, globally and more recently (between June 13 and July 21, 2025) which means using the most recent SOTA models
> I'm not saying the DORA study is more accurate, but at least it surveyed 5000 developers, globally and more recently
It's asking a completely different question; it is a survey of peoples' _perceptions of their own productivity_. That's basically useless; people are notoriously bad at self-evaluating things like that.
It didn't "survey" devs. It paid them to complete real tasks while they were randomly assigned to use AI or not, and measured the actual time taken to complete the tasks vs. just the perception. It is much higher quality evidence than a convenience sample of developers who just report their perceptions.
https://www.augmentcode.com/product/intent
can use the code AUGGIE to skip the queue. Bring your own agent (powered by codex, CC, etc) coming to it next week.