Weaviate

circle-info

Weaviate is an AI-native, open-source vector database designed for semantic search, hybrid search, and RAG (Retrieval-Augmented Generation) applications. It stores both objects and their vector embeddings and supports built-in ML model integration.

Overview

Weaviate goes beyond traditional vector databases by natively integrating ML models for automatic vectorization at import and query time. It supports multiple data types (text, images, video, audio), built-in hybrid search combining BM25 and vector similarity, and multi-tenant deployments. Weaviate is production-ready, cloud-native, and designed to scale from prototypes to billions of vectors.

Property
Value

Category

Vector Database / RAG Infrastructure

Developer

Weaviate B.V.

License

BSD 3-Clause

Stars

12K+

Docker Image

cr.weaviate.io/semitechnologies/weaviate

Ports

22 (SSH), 8080 (HTTP API / GraphQL)


Key Features

  • Vector + keyword hybrid search — combine BM25 full-text with vector similarity in one query

  • Built-in vectorizers — auto-vectorize data at import with OpenAI, Cohere, HuggingFace, or local models

  • Multi-modal — store and search text, images, video, audio in one database

  • GraphQL API — expressive query language for complex semantic queries

  • REST API — full CRUD operations and schema management

  • Multi-tenancy — isolate data per tenant with shared infrastructure

  • HNSW indexing — fast approximate nearest-neighbor search

  • Filtered search — combine vector search with traditional metadata filters

  • Generative search — built-in RAG with LLM integration

  • Horizontal scaling — shard and replicate across multiple nodes

  • Modules system — plug in vectorizers, readers, generators


Clore.ai Setup

Step 1 — Choose Hardware

Use Case
Recommended
RAM
Storage

Development / prototyping

CPU instance

8 GB

20 GB

Small production (< 1M vectors)

CPU instance

16 GB

50 GB

Large scale (10M+ vectors)

GPU instance

32 GB+

200 GB+

GPU-accelerated vectorization

RTX 4090

24 GB

100 GB

circle-info

Weaviate itself runs on CPU. Use GPU instances on Clore.ai when you need local embedding model inference (e.g., text2vec-transformers with a local model) for fast vectorization at import time.

Step 2 — Rent a Server on Clore.ai

  1. Go to clore.aiarrow-up-rightMarketplace

  2. For pure vector search: CPU instances with ≥ 16 GB RAM

  3. For GPU-accelerated embeddings: RTX 3090 or 4090

  4. Open ports: 22 and 8080

  5. Ensure ≥ 50 GB disk for vector storage

Step 3 — Deploy with Docker

Minimal deployment (no vectorizer):

With OpenAI vectorizer:

With local HuggingFace vectorizer (GPU-accelerated):

Start:


Accessing the API

HTTP/REST API

GraphQL Endpoint

Health Check

Via SSH


Python Client

Installation

Connect


Schema & Collections

Create a Collection


Importing Data

Batch Import with Pre-computed Vectors

Auto-vectorize with OpenAI (at import)


Querying

Hybrid Search (Vector + BM25)

Keyword Search (BM25)

GraphQL Query


Generative Search (RAG)


Multi-Tenancy


REST API Examples


Troubleshooting

circle-exclamation
circle-exclamation
circle-info

High memory usage — Weaviate keeps vector index in RAM for fast search. For 1M 768-dim vectors: ~6 GB RAM. Plan accordingly when choosing Clore.ai instance size.

circle-info

Cannot connect via Python client — Ensure both port 8080 (HTTP) and port 50051 (gRPC) are open. The v4 Python client uses gRPC by default.

Issue
Fix

Connection refused

Wait for startup (~30 sec), check docker ps, verify ports

Schema already exists

Delete collection first: client.collections.delete("Name")

Out of memory

Increase RAM or reduce vector dimensions

Slow vector search

Add HNSW index or check dataset size vs available RAM


Performance Tips

  1. Use batch imports — 10x–50x faster than single inserts

  2. Choose right embedding modelall-MiniLM-L6-v2 (384 dims) is fast; text-embedding-3-large (3072 dims) is best quality but uses 8x more RAM

  3. Hybrid search alpha — tune alpha for your use case: 0.25 for keyword-heavy queries, 0.75 for semantic queries

  4. HNSW parametersef and efConstruction control recall vs. speed tradeoff

  5. Tenant isolation — use multi-tenancy for SaaS apps; it scales much better than separate collections per user


  • Qdrant — Rust-based vector database with payload filters

  • ChromaDB — lightweight embeddings database

  • Milvus — high-scale vector database


Weaviate on Clore.ai gives you a production-grade vector database with GPU-accelerated vectorization — ideal for building scalable RAG systems and semantic search applications.


Clore.ai GPU Recommendations

Use Case
Recommended GPU
Est. Cost on Clore.ai

Development/Testing

RTX 3090 (24GB)

~$0.12/gpu/hr

Production Vector Search

RTX 3090 (24GB)

~$0.12/gpu/hr

High-throughput Embedding

RTX 4090 (24GB)

~$0.70/gpu/hr

💡 All examples in this guide can be deployed on Clore.aiarrow-up-right GPU servers. Browse available GPUs and rent by the hour — no commitments, full root access.

Last updated

Was this helpful?