Stop Approving ls: Using a Local LLM to Auto-Classify Command Safety
If you use AI coding assistants like Cline, Cursor, or Claude Code, you know this pain: [y/n/s/a]: [y/n/s/a]: y
A three-part series about building robust consent systems for AI coding assistants. Each post documents a real security vulnerability I discovered while working with Claude, and how we caught and fixed it together.
If you use AI coding assistants like Cline, Cursor, or Claude Code, you know this pain: [y/n/s/a]: [y/n/s/a]: y
Last week I published about building a local LLM command safety classifier. I thought I had command approval figured out. Then my AI assistant got sneaky.
Part 3 of the AI Consent Security series. Previously: Local LLM Command Safety and Trusted Commands Betrayal.
When you give an AI assistant access to your terminal, how do you keep it from running dangerous commands? This series chronicles my journey building a consent system that went from naive allowlisting to sophisticated command analysis.
What makes this unique: Each vulnerability was discovered during actual development work. The AI (Claude) helped me identify and fix its own potential exploits.
cat >> betrayal)python /tmp/evil.py)✅ Series Complete - All 3 parts published!
Human-AI security collaboration in action.