# Inference & Deployment

- [Deploying a REST API for Model Inference](https://dev.clore.ai/inference-and-deployment/rest-api-deployment.md)
- [Building Inference-as-a-Service with FastAPI](https://dev.clore.ai/inference-and-deployment/fastapi-inference.md)
- [Batch Inference at Scale: Processing 1M Images](https://dev.clore.ai/inference-and-deployment/batch-inference.md)
- [Real-Time Video Processing Pipeline](https://dev.clore.ai/inference-and-deployment/video-processing.md)
- [Multi-Model Inference Router](https://dev.clore.ai/inference-and-deployment/model-router.md)
- [Auto-Scaling Inference Workers](https://dev.clore.ai/inference-and-deployment/auto-scaling-workers.md)
