Batch Inference at Scale: Processing 1M Images

What We're Building

A high-throughput batch inference pipeline that processes millions of images using Clore.ai GPUs. Automatically provisions workers, distributes workload, and handles failures — all optimized for cost and speed.

Key Features:

  • Process 1M+ images efficiently

  • Multi-GPU parallel processing

  • Automatic load balancing

  • Checkpoint and resume support

  • S3/GCS integration

  • Cost tracking per batch

  • Progress monitoring

Prerequisites

  • Clore.ai account with API key

  • Python 3.10+

  • Image dataset (local or cloud storage)

pip install requests paramiko scp boto3 tqdm pillow torch torchvision

Architecture Overview

Full Script: Batch Inference System

Performance Benchmarks

Setup
Images
Time
Cost
Images/sec

1x RTX 3080

100K

25 min

$0.13

67

3x RTX 3080

100K

9 min

$0.14

185

5x RTX 4090

1M

45 min

$1.88

370

10x RTX 4090

1M

25 min

$2.08

667

Cost Comparison

Provider
1M Images
Time
Cost

Clore.ai (5x RTX 4090)

1M

45 min

$1.88

AWS Lambda

1M

3 hours

$15.00

GCP Cloud Run

1M

2 hours

$12.00

Local RTX 4090

1M

4 hours

~$1.00 (power)

Next Steps

Last updated

Was this helpful?