A free, Claude-like agent
Local models + skills + a dash of commercial LLM
Build a free, Claude-like experience that anyone can run at home. We achieve this by mixing three powerful layers:
Gemma / Qwen via Ollama handle planning and execution — no per-token cost, no data leaving your machine.
Turn repeated work into Python functions, eliminating LLM calls after the second or third run.
Commercial model pulled in only when local path genuinely can't finish — timeouts, failures, low confidence.
The result: high capability at a fraction of the cost of calling a frontier model for every single step.
Planning and execution run on your machine via Ollama
Timeouts and failures trigger Claude intervention automatically
Repeated tasks become reusable Python skills, cutting LLM calls
Submit tasks, watch tool calls stream live, manage settings
High-risk shell commands pause for approval before running
ChromaDB-backed long-term memory feeds context into future plans
Powerful capabilities that save you money while enabling unlimited creativity
Track your token usage and see exactly how much you're saving compared to using GPT-4 for every task. The hybrid approach keeps costs low while maintaining high quality.
Automatically generate comprehensive reports from your data. OpenTeddy can analyze, visualize, and create professional documents.
Handle complex development tasks including testing, debugging, and data processing. Skills grow automatically to handle repeated workflows.
low confidence | timeout | failure signal | unhealthy
Intel / Apple Silicon - Fully supported
Any distro with Python 3.11+ and Ollama
Partial support - WSL2 recommended
Fully supported - behaves like Linux
OpenTeddy is a solo side-project trying to prove that a small open stack can get close to the big commercial agents.
The single biggest encouragement for the project
Report bugs or share feedback
Contribute your custom skills via PR