Claw Dev is a local multi-provider coding assistant launcher for the bundled terminal client in this repository. It provides a single entry point to choose how model requests are resolved at startup, supporting various AI providers.
- Flexible AI Provider Selection: Choose between Anthropic, Google Gemini, Groq, or Ollama directly at startup.
- Anthropic-Compatible Proxy: Seamlessly integrate Gemini, Groq, and Ollama through a local proxy, maintaining a consistent terminal experience.
- Consistent User Experience: Designed to feel like one unified tool, with unified launcher, prompts, environment variables, and documentation.
- Local Inference Support: Best-in-class support for Ollama, enabling local inference without relying on cloud API providers.
Claw Dev integrates with various AI models and services, utilizing a Node.js-based client with an internal compatibility proxy.
Ensure you have the following installed:
- Node.js: Version 22 or newer.
- npm: Node Package Manager (comes with Node.js).
- Git for Windows: Recommended for Windows users for optimal terminal workflow.
Provider-specific requirements:
- Anthropic: An Anthropic account for in-app login, or
ANTHROPIC_API_KEY. - Gemini:
GEMINI_API_KEY. - Groq:
GROQ_API_KEY. - Ollama: A running Ollama installation with at least one pulled model (e.g.,
qwen3).
Clone the repository:
git clone https://github.com/akariwill/Claw-Dev.gitFrom the repository root:
cd E:\Claw-Dev # Adjust path if different
npm install
copy .env.example .envNote: Editing .env is optional. Claw Dev can prompt for missing values interactively when it starts.
Start Claw Dev from the repository root:
npm run claw-devAlternatively, launch directly from the bundled client directory:
cd E:\Claw-Dev\Leonxlnx-claude-code # Adjust path if different
.\claw-dev.cmdWhen Claw Dev starts, it presents a provider selector:
- Anthropic
- Gemini
- Groq
- Ollama
If a required API key is missing, Claw Dev will prompt for it.
Claw Dev operates in two primary modes to maintain a consistent terminal experience while supporting diverse model backends:
- Anthropic Mode:
- The bundled client communicates directly with Anthropic APIs.
- Compatibility Mode:
- The bundled client interacts with a local proxy (
src/anthropicCompatProxy.ts). - This local proxy translates Anthropic-style
/v1/messagesrequests into native API calls for Gemini, Groq, or Ollama.
- The bundled client interacts with a local proxy (
Ollama offers local inference, ideal for those who prefer not to depend on cloud API providers.
Install Ollama from its official download page:
Ensure the Ollama application or service is running post-installation.
For a quick start, pull a lightweight model:
ollama pull qwen3Verify model availability:
ollama listIf not already running in the background:
ollama serveThe default local API base URL is http://127.0.0.1:11434.
cd E:\Claw-Dev # Adjust path if different
npm run claw-devThen choose option 4. Ollama.
Claw Dev will route requests through its local compatibility proxy to your Ollama server.
Preconfigure Ollama mode in your .env file:
OLLAMA_BASE_URL=http://127.0.0.1:11434
OLLAMA_MODEL=qwen3
OLLAMA_API_KEY= # Not required for local Ollama on localhost
OLLAMA_KEEP_ALIVE=30m # Keeps model loaded, reduces warm-up time
OLLAMA_NUM_CTX=2048 # Controls prompt context size
OLLAMA_NUM_PREDICT=128 # Limits output length, can reduce latencyTo check which models are loaded and their processor usage:
ollama psTo confirm the Claw Dev proxy health:
npm run proxy:compatThen open http://127.0.0.1:8789/health in your browser. A JSON response with the active provider and model should appear when Ollama mode is configured.
Consider these points for optimal performance:
- Larger context windows and longer outputs generally lead to slower responses.
- First-token latency is usually highest on the initial request after a model loads.
- CPU-only inference is significantly slower than GPU-backed inference.
Recommended starting values for responsiveness:
OLLAMA_KEEP_ALIVE=30m
OLLAMA_NUM_CTX=2048
OLLAMA_NUM_PREDICT=128Adjust OLLAMA_NUM_CTX for quality vs. speed. Reduce OLLAMA_NUM_PREDICT for shorter answers and lower latency. If ollama ps shows 100% CPU, slow generation is expected; consider smaller models, optimizing OLLAMA_NUM_CTX, OLLAMA_NUM_PREDICT, and leveraging OLLAMA_KEEP_ALIVE.
Configure your .env file with the appropriate API keys and model names:
ANTHROPIC_API_KEY=your_anthropic_api_key_here
ANTHROPIC_MODEL=claude-sonnet-4-20250514GEMINI_API_KEY=your_gemini_api_key_here
GEMINI_MODEL=gemini-2.5-flashGROQ_API_KEY=your_groq_api_key_here
GROQ_MODEL=openai/gpt-oss-20bOLLAMA_BASE_URL=http://127.0.0.1:11434
OLLAMA_MODEL=qwen3
OLLAMA_API_KEY=
OLLAMA_KEEP_ALIVE=30m
OLLAMA_NUM_CTX=2048
OLLAMA_NUM_PREDICT=128cd E:\myclaudecode\Leonxlnx-claude-code # Adjust path if different
.\claw-dev.cmd --versionSkip the provider menu:
.\claw-dev.cmd --provider anthropic
.\claw-dev.cmd --provider gemini
.\claw-dev.cmd --provider groq
.\claw-dev.cmd --provider ollamaLegacy aliases (--provider claude, --provider grok) are also supported.
echo "Summarize this repository" | .\claw-dev.cmd --bare -pclaw-dev/
├── claude-code/ # Bundled terminal client and Windows launchers
│ └── ...
├── src/anthropicCompatProxy.ts # Local Anthropic-compatible proxy for Gemini, Groq, Ollama
├── .env.example # Optional environment template for local setup
├── package.json # Root scripts for launching, building, and validating
└── ... # Other project files
Before public commits, verify your local Git identity:
Recommended settings:
git config user.name "YOURUSERNAME"
git config user.email "YOUREMAIL"Verify active values:
git config user.name
git config user.emailImportant notes:
.env,node_modules,dist, and*.logfiles are ignored by.gitignore.- Always review
git statusbefore staging andgit diff --cachedbefore pushing.
- Ensure Ollama is installed and its service/app is running.
- Confirm
ollama serveis active if required. - Verify the selected model was pulled successfully.
- Check that
OLLAMA_BASE_URLin.envpoints to the correct server.
Common causes include CPU-only inference, excessively large models for your hardware, context windows that are too large, or long requested outputs.
Use ollama ps to inspect model loading. If PROCESSOR shows 100% CPU, slow generation is expected. Consider smaller models, optimizing OLLAMA_NUM_CTX, OLLAMA_NUM_PREDICT, and leveraging OLLAMA_KEEP_ALIVE.
This indicates Claw Dev is likely functioning, but your local Ollama server might be unreachable or lacks the requested model.
For the shortest setup path when sharing this repository:
- Install Node.js 22 or newer.
- Run
npm install. - Start
npm run claw-dev. - Choose a provider.
- Supply credentials or run Ollama locally.
A separate global installation of the bundled client is not required.
Run these commands for useful checks:
npm run check
npm run build
npm run claw-dev -- --versionOfficial documentation and resources used for this setup:
- Ollama Documentation
- Ollama API Introduction
- Ollama API Authentication
- Ollama FAQ
- Anthropic Claude Code Quickstart
- Groq Docs
This project is licensed under the MIT License.
If you have any questions or feedback, feel free to contact me on Discord wildanjr_ or Instagram akariwill.
