

Something like Stackoverflow is probably the biggest source of code to train a LLM on, and since it’s based around posting code that almost works but you got some problem with it, I’m absolutely not surprised that the LLMs would pick up the habit of making the same subtle small mistakes that humans make.






We were having pretty big problems with our yarn configurations a few weeks ago because the default npm registry is now blocked for security reasons and we’re supposed to only use our own internal registry. The coworker who was supposed to fix it asked his AI whether he can override the configuration for the location for the default registry and the AI said no, so he closed the ticket as “can’t do anything about it I guess 🤷”. A few days later, another coworker just went and changed that configuration by looking at how custom scopes are defined and using common sense. Worked instantly, we haven’t been having any yarn issues since.