> LLMs only reliably know what you just told them, don't rely on training data
This depends a lot on the model but i've been using 4o for all kinds of information retrieval and found it to be generally reliable. At least not much worse than the general internet. You can ask it for sources and of course you should not use it as an authority, but it can often be a very good way to quickly find out a fact. You do need to developed a feeling for the kind of thing it will know realiably and the kind of thing that will cause it to start halucinating (a bit like some of your co-workers).
> LLMs cannot write for you
I disagree, LLMs can write small blocks of text very well. But there is an art to using it. Don't try to create too much at once. I often find it works better then I give less input. If you list a bunch of things it needs to include, it tends the result reads like a student trying to include all the buzzwords.
> LLMs can help a human perform tasks, they cannot replace a human
I don't think anybody claims otherwise for current public models.
> Have the LLM do as little as possible
You need to learn what LLMs do well, and then use it for that. The idea that it is most efficient to program everything by hand as much as possible does not match my experience. Writing boilerplate code of under 50 lines of code or so is something current models already do very well and very quickly. I usually just try to generate it and if it does not work I write it by hand.
Finally, LLMs now take video and audio. We use Gemini to write meeting notes from Google meet and they tend to be very high quality, a lot better then what a random person taking notes usually produces. So the models are not text-only.
ChatGPT with search is an example of RAG, which is a pattern this article is promoting: better results from the LLM because ChatGPT injected additional search result context into the model to accompany your question.
I think of contradictory examples where the article doesn't make sense: LLM Chat products are just the product of training data. Generative coding applications take 1 prompt and generate a lot of code.
What this article did make me think is the existing Chat UI in coding apps are too limiting. Some have image attachments, but we need to allow users to input more detail in their prompt (visually, or by having a pre-generation discussion about specifics). That's why I think product engineers will benefit from AI more than non technical folk.
Also, do you have resources that align with those opinions?
Example: Should we use "AI" for authentication and authorization?
For logging in: No.
For checking authority for an operation: No.
For determining the likelihood that the IP address a login attempt is coming from is part of an attack pattern: Yes!
Is what you're doing taking a large amount of text and asking the LLM to convert it into a smaller amount of text? Then it's probably going to be great at it. If you're asking it to convert into a roughly equal amount of text it will be so-so. If you're asking it to create more text than you gave it, forget about it.
For me, I think the one exception to that is code: I can often get a few dozen lines of working code from the right single sentence prompt.
I can get better code if I feed in more examples of the libraries I'm using though, which fits this more-input-is-better rule.
I use chatgpt like a chat-with-wikipedia. For general knowledge where I was too lazy to google or which is too far hidden in ads ridden adsense blogs, like "give me an easy recipe for waffles"
Here it becomes interesting: I have been coding with Aider and sonnet3.5 for the past 3-4 weeks. I add the few files I need to change as well as whatever structural information (db-scheme or so) and ask it to work on a task, very high level. Yes, I need to offer architectural advice or if I want library X or Y included, but overall, it produces very good (I would say junior level) code.
The thing is, it does all the long writing I got tired of over the years. The writing I don't want to do anymore. The CRUD methods. The "change the labels, and names of this to that". The "write a db connector that does XYZ". All the stuff I would normally ask a junior to implement.
The output code is generally ok. It's not an eye candy. But it does work generally, maybe with occasional hiccups.
But the time saved by prompting it with 2-3 lines and then getting code that normally takes me 30 minutes to convince myself that I really need to do that and then 10 minutes to write is now done in mere 20 seconds. Without much of a mental context switch and without being very taxing on the brain.
I am not sure if that counts as "take large amount and make it smaller" based on the extensive code it ingested during training, or if it falls in the category "create more text than I gave it", but it works for me. I would miss it if it didn't work anymore.
"Great article that contains some pithy aphorisms that I expect to see again and again."
"Great article that is approachable enough to share with less technical folk. Thank you!"
"Good article, I would add the good things about working with media different than text. For example, describing images."
They are similar indeed. The users have been around for a while so unless their accounts have been taken over, I wouldn't worry.
Sheesh. I think an AI would write a better article...
I've seen this behaviour in Claude but don't remember 4o doing the the same or as frequently.
You still have 4 people that have been let go who will need to find a way to earn a living in this competitive market.
Going off on a tangent here, but ChatGPT Search is miles ahead of the, with all due respect, garbage that Google Search spews these days. Honestly, Google Search is now in a very awkward position where using an old-style keyword search doesn't really do what I want and asking it a question ChatGPT-style doesn't do what I want, either. I don't even know how to use it anymore to get good results. And why even bother when ChatGPT Search exists?