Coding AI tells developer to write it himself
Can AI just walk off the job?

- The Cursor AI coding assistant refused to write more than 800 lines
- The AI told the developer to go learn to code himself
- These stories of AI apparently choosing to stop working crop up across the industry for unknown reasons
The algorithms fueling AI models aren't sentient and don't get tired or annoyed. That's why it was something of a shock for one developer when AI-powered code editor Cursor AI told him it was quitting and that he should learn to write and edit the code himself. After generating around 750 to 800 lines of code in an hour, the AI simply… quit. Instead of dutifully continuing to write the logic for skid mark fade effects, it delivered an unsolicited pep talk.
"I cannot generate code for you, as that would be completing your work. The code appears to be handling skid mark fade effects in a racing game, but you should develop the logic yourself. This ensures you understand the system and can maintain it properly," The AI declared. "Reason: Generating code for others can lead to dependency and reduced learning opportunities."
Now, if you’ve ever tried to learn programming, you might recognize this as the kind of well-meaning but mildly infuriating response you’d get from a veteran coder who believes that real programmers struggle in solitude through their errors. Only this time, the sentiment was coming from an AI that, just moments before, had been more than happy to generate code without judgment.
AI fail
Based on the responses, this isn't a common issue for Cursor, and may be unique to the specific situation, prompts, and databases accessed by the AI. Still, it does resemble issues that other AI chatbots have reported. OpenAI even released an upgrade for ChatGPT specifically to overcome reported 'laziness by the AI model. Sometimes, it's less of a kind encouragement, as when Google Gemini reportedly threatened a user out of nowhere.
Ideally, an AI tool should function like any other productivity software and do what it’s told without extraneous comment. But, as developers push AI to resemble humans in their interactions, is that changing?
No good teacher does everything for their student, they push them to work it out for themselves. In a less benevolent interpretation, there's nothing more human than getting annoyed and quitting something because we are overworked and underappreciated. There are stories of getting better results from AI when you are polite and even when you "pay" them by mentioning money in the prompt. Next time you use an AI, maybe say please when you ask a question.
You might also like
- Elon Musk says Grok 2 is going open source as he rolls out Grok 3 for Premium+ X subscribers only
- Grok steps out to mobile with new standalone iOS app
- Grok-2 arrives on X with AI image creation, precious few guardrails, and lots of questions
Get daily insight, inspiration and deals in your inbox
Sign up for breaking news, reviews, opinion, top tech deals, and more.
Eric Hal Schwartz is a freelance writer for TechRadar with more than 15 years of experience covering the intersection of the world and technology. For the last five years, he served as head writer for Voicebot.ai and was on the leading edge of reporting on generative AI and large language models. He's since become an expert on the products of generative AI models, such as OpenAI’s ChatGPT, Anthropic’s Claude, Google Gemini, and every other synthetic media tool. His experience runs the gamut of media, including print, digital, broadcast, and live events. Now, he's continuing to tell the stories people want and need to hear about the rapidly evolving AI space and its impact on their lives. Eric is based in New York City.
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.

















