A lightning-fast search engine API bringing AI-powered hybrid search to your sites and applications.
- Updated
- Rust
![]() |
VOOZH | about |
A lightning-fast search engine API bringing AI-powered hybrid search to your sites and applications.
Qdrant - High-performance, massive-scale Vector Database and Vector Search Engine for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of a cloud-native database.
Simple, Elastic-quality search for Postgres
The AI-native database built for LLM applications, providing incredibly fast hybrid search of dense vector, sparse vector, tensor (multi-vector), and full-text.
Distributed vector search for AI-native applications
A multi-modal vector database that supports upserts and vector queries using unified SQL (MySQL-Compatible) on structured and unstructured data, while meeting the requirements of high concurrency and ultra-low latency.
A Markdown-first memory system, a standalone library for any AI agent. Inspired by OpenClaw.
Lite & Super-fast re-ranking for your search & retrieval pipelines. Supports SoTA Listwise and Pairwise reranking based on LLMs and cross-encoders and more. Created by Prithivi Da, open for PRs & Collaborations.
Endee.io – A high-performance vector database, designed to handle up to 1B vectors on a single node, delivering significant performance gains through optimized indexing and execution. Also available in cloud https://endee.io/
RAG Time: A 5-week Learning Journey to Mastering RAG
Embeddable, in-memory, document-oriented database with a high-level Query builder interface.
The codebase for the book "AI-Powered Search" (Manning Publications, 2025)
A cutting-edge search engine project tailored specifically for the AI product
A LLM RAG system runs on your laptop. 大模型检索增强生成系统,可以轻松部署在笔记本电脑上,实现本地知识库智能问答。
Official Repo of Moss
Local-first RAG server for developers. Semantic + keyword search for code and technical docs. Works with MCP or CLI. Fully private, zero setup.
Coltt is a vector database that supports Multi-Vector Search, high-performance HNSW, FLAT and quantization, and enables fast searches through sophisticated internal data shard design.
Production-ready RAG Framework (Python/FastAPI). 1-line config swaps: 6 Vector DBs (Weaviate, Pinecone, Qdrant, ChromaDB, pgvector, MongoDB), 5 LLMs (Gemini, OpenAI, Claude, Ollama, OpenRouter). OpenAI-compatible API. 2100+ tests.
DocMind AI is a powerful, open-source Streamlit application leveraging LlamaIndex, LangGraph, and local Large Language Models (LLMs) via Ollama, LMStudio, llama.cpp, or vLLM for advanced document analysis. Analyze, summarize, and extract insights from a wide array of file formats, securely and privately, all offline.
Add a description, image, and links to the hybrid-search topic page so that developers can more easily learn about it.
To associate your repository with the hybrid-search topic, visit your repo's landing page and select "manage topics."