DeepSeek Coder

Best-in-class code generation with DeepSeek Coder models.

circle-check

Renting on CLORE.AI

  1. Filter by GPU type, VRAM, and price

  2. Choose On-Demand (fixed rate) or Spot (bid price)

  3. Configure your order:

    • Select Docker image

    • Set ports (TCP for SSH, HTTP for web UIs)

    • Add environment variables if needed

    • Enter startup command

  4. Select payment: CLORE, BTC, or USDT/USDC

  5. Create order and wait for deployment

Access Your Server

  • Find connection details in My Orders

  • Web interfaces: Use the HTTP port URL

  • SSH: ssh -p <port> root@<proxy-address>

What is DeepSeek Coder?

DeepSeek Coder offers:

  • State-of-the-art code generation

  • 338 programming languages

  • Fill-in-the-middle support

  • Repository-level understanding

Model Variants

Model
Parameters
VRAM
Context

DeepSeek-Coder-1.3B

1.3B

3GB

16K

DeepSeek-Coder-6.7B

6.7B

8GB

16K

DeepSeek-Coder-33B

33B

40GB

16K

DeepSeek-Coder-V2

16B/236B

20GB+

128K

Quick Deploy

Docker Image:

Ports:

Command:

Accessing Your Service

After deployment, find your http_pub URL in My Orders:

  1. Go to My Orders page

  2. Click on your order

  3. Find the http_pub URL (e.g., abc123.clorecloud.net)

Use https://YOUR_HTTP_PUB_URL instead of localhost in examples below.

Using Ollama

Installation

Code Generation

Fill-in-the-Middle (FIM)

DeepSeek-Coder-V2

Latest and most powerful:

vLLM Server

API Usage

Code Review

Focus on:

  1. Performance

  2. Readability

  3. Best practices """} ]

"""} ]

Performance

Model
GPU
Tokens/sec

DeepSeek-1.3B

RTX 3060

~120

DeepSeek-6.7B

RTX 3090

~70

DeepSeek-6.7B

RTX 4090

~100

DeepSeek-33B

A100

~40

DeepSeek-V2-Lite

RTX 4090

~50

Comparison

Model
HumanEval
Code Quality

DeepSeek-Coder-33B

79.3%

Excellent

CodeLlama-34B

53.7%

Good

GPT-3.5-Turbo

72.6%

Good

Troubleshooting

Code completion not working

  • Ensure correct prompt format with <|fim_prefix|>, <|fim_suffix|>, <|fim_middle|>

  • Set appropriate max_new_tokens for code generation

Model outputs garbage

  • Check model is fully downloaded

  • Verify CUDA is being used: model.device

  • Try lower temperature (0.2-0.5 for code)

Slow inference

  • Use vLLM for 5-10x speedup

  • Enable torch.compile() for transformers

  • Use quantized model for large variants

Import errors

  • Install dependencies: pip install transformers accelerate

  • Update PyTorch to 2.0+

Cost Estimate

Typical CLORE.AI marketplace rates (as of 2024):

GPU
Hourly Rate
Daily Rate
4-Hour Session

RTX 3060

~$0.03

~$0.70

~$0.12

RTX 3090

~$0.06

~$1.50

~$0.25

RTX 4090

~$0.10

~$2.30

~$0.40

A100 40GB

~$0.17

~$4.00

~$0.70

A100 80GB

~$0.25

~$6.00

~$1.00

Prices vary by provider and demand. Check CLORE.AI Marketplacearrow-up-right for current rates.

Save money:

  • Use Spot market for flexible workloads (often 30-50% cheaper)

  • Pay with CLORE tokens

  • Compare prices across different providers

Next Steps

Last updated

Was this helpful?