What’s interesting here is that with AI, all our interfaces should evolve away from previous generation rigid forms / buttons / tables etc. towards something more fluid / dynamic / “natural”. Yet all the AI coding is geared towards producing more of the former.
Most of the software we interact with is at the end of the day some db tables, queries to read/write, and some ui to read/write. There have been so many times I wished I could just do my own db joins on the underlying db to get the views I wanted. But I can’t - because the app has pre-defined ui/query paths.
With AI, I should be able to ask for things the product designers didn’t anticipate or left out and the system could query, create ui on the fly, etc…
i think of them as tiers of expertise— need to master the basics of structure and form before the robot has the learned representations to competently model user interactions with more fluid instantiations (by downprojecting into the overlearned fixed-semantics)