The relevance for me personally is whether or not they can be useful for programming, and if they’re accessible to run locally. I’m not interested in feeding my data to a datacentre. My AMD GPU also doesn’t support ROCm so LLMs run slow as fuck for me. So, generally, I avoid them.
LLMs consistently produce lower quality, less correct, and less secure code than humans. However, they do seem to be getting better. I might be open to using them to generate unit tests if only they would run faster on my PC. I tried deepseek, llama3.1, and codellama; all take like an hour+ to answer a programming question given that they are just using my CPU, as my GPU doesn’t support ROCm. So really not feasible for anything.
Depending on what you count as AI, I think some of the long-existing predictive ML like autosuggestions based on learning your input patterns are fine and helpful. And maybe if I get a supported GPU I won’t mind using local LLMs for some things. But generally I’m not dying to use them. I can do things myself.





















Over the weekend.
I don’t see small talk or whatever as a sign of a lack of interest in someone. Sometimes there’s nothing that deep going on to talk about.