|
|
--- |
|
|
license: mit |
|
|
--- |
|
|
# gguf-code: IDE AI Coding Assistant |
|
|
|
|
|
Monster coding agent lives in your IDE - moving beyond a CLI bridge to a fully native IDE AI assistant. |
|
|
|
|
|
 |
|
|
|
|
|
Shipped to [vscode marketplace](https://marketplace.visualstudio.com/items?itemName=gguf.gguf-code) and it's compatible IDEs, i.e., cursor, windsurf, antigravity, etc., [marketplaces](https://open-vsx.org/extension/gguf/gguf-code) |
|
|
|
|
|
## Features |
|
|
|
|
|
- **Native Chat Interface**: Deeply integrated webview clean ui for smooth chat interactions inside VS Code (and/or its clone/fork IDEs). |
|
|
- **Embedded AI Engine**: Run Coder's powerful AI logic directly within the extension process - no external CLI tool required. |
|
|
- **Project-Aware Context**: Automatically understands your workspace structure and active files. |
|
|
- **Smart Tools**: |
|
|
- `read_file`: Intelligently reads file contents with progressive loading for large files. |
|
|
- `write_file`, `find_files`, and more (Coming Soon). |
|
|
- **Live Diff Preview**: Preview AI-suggested code changes with standard VS Code diff views. |
|
|
- **Diagnostics Integration**: Uses LSP diagnostics to help fix errors in your code. |
|
|
|
|
|
## Requirements |
|
|
|
|
|
- Visual Studio Code version 1.104.0 or higher |
|
|
- `coder.config.json` in your workspace root (for configuring AI providers). |
|
|
|
|
|
## Getting Started |
|
|
|
|
|
1. **Installation**: Install the extension from the VS Code Marketplace or via `.vsix`. |
|
|
2. **Configuration**: Ensure you have a `coder.config.json` in your project root (see example below, can be any OpenAI compatible API or local/self-hosted API). |
|
|
|
|
|
lmstudio (localhost) example: |
|
|
```json |
|
|
{ |
|
|
"coder": { |
|
|
"providers": [ |
|
|
{ |
|
|
"name": "lmstudio", |
|
|
"models": [ |
|
|
"openai/gpt-oss-20b" |
|
|
], |
|
|
"baseUrl": "http://localhost:1234/v1" |
|
|
} |
|
|
] |
|
|
} |
|
|
} |
|
|
``` |
|
|
|
|
|
ollama (localhost) example: |
|
|
```json |
|
|
{ |
|
|
"coder": { |
|
|
"providers": [ |
|
|
{ |
|
|
"name": "ollama", |
|
|
"models": [ |
|
|
"gpt-oss:20b" |
|
|
], |
|
|
"baseUrl": "http://localhost:11434/v1" |
|
|
} |
|
|
] |
|
|
} |
|
|
} |
|
|
``` |
|
|
|
|
|
fetch from any api provider, see example below: |
|
|
```json |
|
|
{ |
|
|
"coder": { |
|
|
"providers": [ |
|
|
{ |
|
|
"name": "openrouter", |
|
|
"baseUrl": "https://openrouter.ai/api/v1", |
|
|
"apiKey": "your-api-key", |
|
|
"models": ["anthropic/claude-4.5-sonnet"] |
|
|
} |
|
|
] |
|
|
} |
|
|
} |
|
|
``` |
|
|
|
|
|
3. **Launch**: Click the "Code" monster icon in the Activity Bar (left sidebar) to open the chat. |
|
|
|
|
|
## Usage |
|
|
|
|
|
- **Chat**: Type naturally in the chat window. "Explain this file", "Fix this bug", etc. |
|
|
- **Context**: The assistant automatically knows about your open files. |
|
|
|
|
|
## Architecture |
|
|
|
|
|
This extension has migrated from a thin client (CLI bridge) to a thick client architecture: |
|
|
|
|
|
- **Frontend**: React + Vite (Webview). |
|
|
- **Backend**: Node.js Extension Host (running `ai-sdk-client`). |
|
|
- **Communication**: VS Code Message Passing API. |
|
|
|
|
|
## Development |
|
|
|
|
|
```bash |
|
|
npm install |
|
|
|
|
|
# Build both Webview and Extension |
|
|
npm run build |
|
|
|
|
|
# Watch mode (rebuilds on change) |
|
|
|
|
|
# Terminal 1: |
|
|
npm run watch:webview |
|
|
|
|
|
# Terminal 2: |
|
|
npm run watch |
|
|
``` |
|
|
|
|
|
## Contributing |
|
|
|
|
|
Contributions are welcome! |
|
|
|
|
|
## License |
|
|
|
|
|
MIT |
|
|
|