Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I think writing Macbeth isn't a good indicator of anything except being Shakespeare. Asking it a few questions about some general concepts like logic, math, and patterns(relationships in Macbeth maybe?) would probably yield results that are much more likely to pass the test.


I disagree on all counts. The first program that would be generated by such a procedure would very likely be the constant program that happens to always respond with the correct questions to your test. It depends how you randomly generate code, of course, but I doubt the complexity of a "general AI program" would ever be smaller than the complexity of the constant program which by luck returns the correct answer to your test.


Generate the Math tests then? That prevents constants.

Generating macbeth is a prompt most people here couldn’t pass, myself included...


you can still use generating Macbeth as a good indicator, just by providing some inputs and constraints.

Input is a Macbeth outline/play summary (e.g. https://www.cliffsnotes.com/literature/m/macbeth/play-summar... ) Contraints are "in the style of : KingLear and Othello"

and compare the produced text to Macbeth




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: