Hacker Newsnew | past | comments | ask | show | jobs | submit | kaelandt's commentslogin

the AI inference workloads shown in the paper are extremely far from what is implied when one says "... computer for AI inference". No discussion of issues around the memory hierarchy and how the presented architecture solves those. No mention of transformers, except for a vague reference to energy-based models


The API prices of $3/$15 are not right for a lot of models. see at openrouter, the gpt-oss-120b ones https://openrouter.ai/openai/gpt-oss-120b, it's more like $0.01/$0.3 (and that model actually needs h200/b200 to have good throughput).


One thing that is also truly unappreciated is most of us humans actually enjoy thinking, and people are trying to make llms strip us from a fundamental thing we enjoy doing. Look at all the people that enjoy solving problems for the sake of it


could be nice to add a "merged PR with a test" metric. looking at the PRs they are mostly without tests, so could be bogus for all we know


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: