Gemma 2
Run Google's Gemma 2 models efficiently on Clore.ai GPUs
Renting on CLORE.AI
Access Your Server
What is Gemma 2?
Model Variants
Model
Parameters
VRAM
Context
Quick Deploy
Accessing Your Service
Using Ollama
Installation
Basic Usage
Gemma 2 2B (Lightweight)
Gemma 2 27B (Best Quality)
vLLM Server
OpenAI-Compatible API
Streaming
Gradio Interface
Batch Processing
Performance
Model
GPU
Tokens/sec
Comparison
Model
MMLU
Quality
Speed
Troubleshooting
Slow generation
Output quality issues
Tokenizer warnings
Cost Estimate
GPU
Hourly Rate
Daily Rate
4-Hour Session
Next Steps
Last updated
Was this helpful?