CrewAI Multi-Agent Framework

Deploy CrewAI on Clore.ai — orchestrate teams of role-playing autonomous AI agents for complex multi-step tasks using any LLM provider.

Overview

CrewAIarrow-up-right is a cutting-edge framework for orchestrating role-playing autonomous AI agents, with 44K+ GitHub stars. Unlike single-agent systems, CrewAI lets you define specialized agents (Researcher, Writer, Coder, Analyst...) that collaborate as a "crew" to complete complex tasks — each agent with its own role, goal, backstory, and toolkit.

On Clore.ai, CrewAI can be deployed in a Dockerized environment for as little as $0.05–0.20/hr. While CrewAI itself is CPU-bound (it orchestrates API calls), combining it with a local Ollama or vLLM server on the same GPU node gives you a fully private, offline-capable multi-agent system.

Key capabilities:

  • 👥 Multi-agent crews — define agent personas with roles, goals, and backstories

  • 🎯 Task delegation — manager agent automatically assigns tasks to the right specialist

  • 🛠️ Tool ecosystem — web search, file I/O, code execution, database access, custom tools

  • 🔁 Sequential & Parallel — execute tasks in order or run independent tasks simultaneously

  • 🧠 Agent memory — short-term, long-term, entity, and contextual memory types

  • 🔌 LLM-agnostic — works with OpenAI, Anthropic, Google, Ollama, Groq, Azure, and more

  • 📊 CrewAI Studio — visual interface for building crews without code (enterprise)

  • 🚀 Pipelines — chain multiple crews for complex multi-stage workflows


Requirements

CrewAI is a Python library. It runs on CPU and requires only a system Python 3.10+ environment or Docker. GPU is optional but unlocks powerful local model inference.

Configuration
GPU
VRAM
System RAM
Disk
Clore.ai Price

Minimal (cloud APIs)

None / CPU

2 GB

10 GB

~$0.03/hr (CPU)

Standard

None / CPU

4 GB

20 GB

~$0.05/hr

+ Local LLM (small)

RTX 3080

10 GB

8 GB

40 GB

~$0.15/hr

+ Local LLM (large)

RTX 3090 / 4090

24 GB

16 GB

60 GB

$0.20–0.35/hr

+ High-quality local LLM

A100 40 GB

40 GB

32 GB

100 GB

~$0.80/hr

API Keys

CrewAI works with most major LLM providers. You need at least one:

  • OpenAI — GPT-4o (best reasoning for complex tasks)

  • Anthropic — Claude 3.5 Sonnet (excellent for writing-heavy crews)

  • Groq — Free tier, fast inference (Llama 3 70B)

  • Ollama — Fully local, no API key needed (see GPU Acceleration)


Quick Start

1. Rent a Clore.ai server

Log in to clore.aiarrow-up-right:

  • CPU-only if using cloud LLM APIs

  • RTX 3090/4090 for local Ollama inference

  • SSH access enabled

  • No special port requirements for CLI usage (expose ports only for web UIs)

2. Connect and prepare

3. Option A — Direct pip install (fastest)

5. Create your first crew


Configuration

Project structure (from crewai create)

agents.yaml — Define your agents

tasks.yaml — Define tasks

crew.py — Assemble the crew

Running with Docker Compose (with Ollama)


GPU Acceleration

CrewAI itself doesn't use the GPU — but the LLM it calls does. Run Ollama or vLLM on the same Clore server for GPU-accelerated local inference.

Configure CrewAI LLM per agent

Model recommendations for agent tasks

Task Type
Recommended Model
VRAM
Notes

Research + web search

Llama 3.1 70B

40 GB

Best local reasoning

Code generation

Codestral 22B

13 GB

Code-specialized

Writing

Llama 3.1 8B

6 GB

Fast, good quality

Complex orchestration

GPT-4o (API)

Best overall

Embeddings/memory

nomic-embed-text

< 1 GB

Required for memory

See Ollama on Clore.ai and vLLM on Clore.ai for full inference setup guides.


Tips & Best Practices

Cost optimization

Running crews as a persistent service

Useful built-in CrewAI tools

Implementing human-in-the-loop


Troubleshooting

"openai.AuthenticationError" even with valid key

Agent stuck in reasoning loop

CrewAI tools fail (SerperDevTool 403)

Memory errors (ChromaDB / embeddings)

Docker build fails on ARM/x86 mismatch

Rate limiting from LLM APIs


Further Reading

Last updated

Was this helpful?