Creating a new endpoint that accepts data as sent by the AI-generated code is very different from accepting AI-generated errors for existing endpoints. You're talking about the latter.
> I'm a Data Scientist currently consulting for a project in the Real Estate space (utilizing LLMs).
Consultants are obviously making huge amounts of money implementing LLMs for companies. The question is whether the company profits from it afterwards.
Time will tell, but I would cautiously say say yes.
Note that I don't usually work in that particular space (I prefer simple solutions and don't follow the hype), didn't sell myself using 'AI' (I was referred), and also would always tell a client if I believe there isn't much sense in a particular ask.
This particular project really uniquely benefits from this technology and would be much harder, if possible at all, otherwise.
Not GP, but I would imagine "another checker to scan the results" would be another NN classifier.
Thinking being that you'd compare outputs of the two, and under assumption of the results being statistically independent from each other and of similar quality, say 1% difference between the two in said comparison, would suggest ~ 0.5% error rate from "ground truth".
I've never understood the fascination with programming languages among computer science folks. Just write machine code directly, it's what it compiles to anyway.