I was at conference a few months ago and, of course, AI was a big topic. The joke was made that nobody knows what to do with it yet, but everybody is demanding that for every product we have to go ahead and take out all the blockchain and stuff it full of AI.
At their hearts, the large language models are statistical words algorithms. Yes, there is more going on there, but what they are good at is computing the statistically most probably word that would follow the previous word. And they are actually quite good at that. The problem is that we, as humans, are very susceptible to believing wrongness that is delivered with high confidence.
Here's an analogy: Bob is a physicist who has a PhD in whatever physicists overeducate themselves on and knows everything there is to know about how light particles move through cardboard and, since it's my story, his theories are all 100% correct and accurate. But he has so much information on the topic that when he is asked to give a 10 minute speech to a kindergarten class, it's not going to turn out well. Fred is a speechwriter who is trained and practiced and amazing at his job and by working with the speaker can produce amazingly easy to understand speeches for audiences of any level. Fred will sit down with Bob and pick through his brain to distill the physics into a speech that the kindergarten class will enjoy and understand. Fred doesn't know E36 M3 about physics, though. So if Fred is told to also incorporate information from Sam, but Sam is a crackpot who has completely wrong theories, Fred doesn't know enough to be able to give weight to one expert or the other. He'll just string the words together so they sound good.
The other problem that AI LLM has is that it is susceptible to hallucinations. In my view, the most dangerous of those are when the system has to make inferences because it can't find an answer and that is driven by the programming and design of the LLM. For example, I asked it for the source of a paraphrased quote that I remember from a magazine article a while ago. It couldn't find it, but it very confidently told me that it was attributed to Steve Jobs. Now, I was pretty sure that was wrong, so I asked it who attributed it to Steve Jobs. It got a little squirrelly and changed its mind and said that it thought that it sounded like something he would have said. I pressed it further and it finally came back and said that it was sorry, but it actually couldn't find anything about that quote at all.
Somewhere in there, they need to program the LLM to be able to fill in gaps and make inferences. When they do that, there are decisions that have to be made about how aggressively it is going to fill in gaps - if it didn't fill them in at all, it would be almost impossible to get any information at all from them.
I have been using both copilot and chatgpt for help with a software development project I've been working on as well as with doing some resume work and it has been pretty good with those.
For the coding help, it's awesome at being able to debug things that would take me a long time to find. For example, I was trying to output an HTML table as a PDF using a specific module and while the HTML would render properly in a browser, the PDF wouldn't render at all. I stared at it for a while and couldn't find the problem. When I fed the code into chatgpt, however, it instantly noticed that I had a typo in an HTML tag. The browser just handled it, but the PDF module had a silent choke on it.
For resume and writing help, I've used it to prompt me to write. For example, I had it walk me through the process of developing my STAR stories by getting it to ask me the questions I needed to answer. And then, I could type my stream of consciousness back in and it would structure it for me into bullet points that I would then use to write paragraphs. Same stuff that I usually do for myself when I write, but it speeds up the process a bit.
What it isn't good at is validating it's own training data. It can't make a value judgement, it just knows what it's training data usually does. That's how you have LLMs rickrolling people. It doesn't know WHY people are always answering questions with a link to this YouTube video, but it happens so much that it must be the statistically most likely thing that should be sent.
What's going to get wicked, though, is when adoption of agentic AI starts to snowball. These are AI models that don't string words together but that actually perform an action. For example, imagine an agentic AI that learns that the most statistically likely thing for it to do when it senses that it is 40 degrees outside, the motion sensors in the house are quiet, and the left garage door begins to open is to turn on the lights in the living room, raise the temperature in the house to 76 degrees, turn on the fireplace, and turn on the TV and start the F1 TV app. Or, in a business sense, when it detects that the CFO just sent an email to someone they've never communicated with in the shipping department and there's some mention about sending payments, it will block the email, alert security, and initiate a lock on the last few snapshots of data in case a restore has to happen.