Discussion about this post

User's avatar
Alexey Gumirov's avatar

Regaring your proposal to "Use Interpretive Agents When Ambiguity Exists". The harderst point for me is to verify how agent does the job, its outcomes. How can we verify what is good and bad, right or wrong in the ambiguous world? If I want to build and automaton which is doing some task at scale, I need some metric(s) to measure its performance and quality. And I also need to understand how can I improve its quality at scale?

E.g. for the deterministic program I know, that for 100 calls per day accuracy of 99.9% gives me about 1 error per 10 days of work. If I need to scale up to 1000 calls per day and keep the same 1 error per 10 days of work, I have to increase its accuracy to 99,99%. I most probably can do it because it is deterministic program, I know where to look at what to do in order to improve it.

But with LLM (AI) based agent it is not the case. Even with the fine tuning or re-training you never sure. Unless agent context is narrowed down before it reaches the model, so that probability of drifting/hallucination is reduced. But eventually at peak scale we end up with very very narrow trained AI inferences, surrounded by the deterministic pre-filters, data transformers, input ambigiuty cleaning programs (most probably deterministic). And in this case the question will be - do I need AI here at all, if I already transformed my data so that I most probably achieve better results with just deterministic program in the end?

Alexey Gumirov's avatar

Thanks for a long detailed article! But I would argue with some statements there:

"An LLM can execute deterministic logic".

LLM can try to make pattern matching for the words which represent deterministic logic. But can it really execute it?

Recent LLM tests on obscured programming languages demonstrated near 0 success rate of LLMs. This research also breaks another of your premises you state many times in your article: "The model can understand context..."

The model has no "understanding" in the human meaning, it is not building absractions, it does not have "the world model". It is just a huge C-written program multiplying tensors with different coefficients. It does it over and over for every new token to produce for find the best (most probable) next pattern.

Understanding means not only being able to replicate, but reproduce something equivalent or even new in the new form, language, art, creation.... Human understanding means to extend, improve your world model. It is actually part of learning.

LLMs are fixed programs, every iteration of running does not lead to self-learning or self-improvement.

This is why quite often I personally feel frustrated and exhausted when I use LLMs. It is like explaining the same thing over and over again to an assistant who never learns.

2 more comments...

No posts

Ready for more?