Comment In my experience (Score 5, Insightful) 56
LLMs are not good at self-management or judgement-call making. Allowing them to be "agents" and do things on your behalf is problematic because they can get things wrong and then make things worse when they try to fix it. They are much worse about this than human agents.
In my experience so far LLMs can generate code that "looks right", but doesn't necessarily work right. The more details there are in the requirements, the worse the LLM does. And in my experience implementing business workflow pipelines using LLMs, the LLMs are pretty good at interpreting plain English requests and translating them to something machine-parseable (like JSON or whatever), so you can then write your own code that reliably takes action, using the LLM just as a bridge between the two. But the more you ask the LLM to solve problems itself, make decisions itself, or take actions itself, the more it lets you down.
So, I think that AI just isn't ready for what Microsoft plans to use it for. And it seems like many others agree.