Continue.dev AI Coding
Power Continue.dev with Clore.ai GPUs — run CodeLlama 34B, DeepSeek Coder, and Qwen2.5-Coder locally on cheap GPU rentals for private AI coding assistance.
Overview
Property
Details
Recommended Models for Coding
Model
VRAM
Strength
Notes
Requirements
Clore.ai Server Requirements
Tier
GPU
VRAM
RAM
Disk
Price
Models
Local Requirements (Your Machine)
Quick Start
Part 1: Set Up the Clore.ai Backend
Option A — Ollama Backend (Recommended for Most Users)
Option B — vLLM Backend (High-Throughput / OpenAI-Compatible)
Option C — TabbyML Backend (FIM Autocomplete Specialist)
Part 2: Install Continue.dev Extension
Part 3: Configure Continue.dev to Use Clore.ai
Configuration
SSH Tunnel Setup (Secure Remote Access)
Persistent Tunnel with autossh
Load Multiple Models for Different Tasks
Codebase Indexing (RAG for Your Repo)
GPU Acceleration
Monitor Inference Performance
Expected Performance by GPU
GPU
Model
Context
Tokens/sec (approx.)
Tune Ollama for Better Performance
Tips & Best Practices
Use Different Models for Different Tasks
Cost Comparison
Solution
Monthly Cost (8hr/day usage)
Privacy
Model Quality
Shut Down When Not Coding
Use Continue.dev Custom Commands
Troubleshooting
Problem
Likely Cause
Solution
Debug Commands
Continue.dev Config Validation
Further Reading
Last updated
Was this helpful?