"The Supabase for Search." — CEO Ghita Houir Alami
Most AI products lie to you — not because the model is wrong, but because it retrieved the wrong thing. ZeroEntropy fixes that. Their rerankers, embeddings, and search infrastructure are the unsexy, load-bearing wall that keeps AI from hallucinating its way into your legal contract or medical record.
Three panels. One uncomfortable truth.
Most teams stitch together a vector database, a keyword search, and a re-ranker from three different vendors. Then maintain it. Then watch it break at 2am.
Others dump the entire knowledge base into the LLM context window. Quick fix. Compounding errors. Expensive. The model confidently answers the wrong question.
One API. Ingestion, indexing, reranking, evaluation. Deploy human-level search in an afternoon. Fewer hallucinations. Faster results. Actually works.
A Moroccan mathematician who left home at 17, and a CMU dropout who audited blockchains for hedge funds. Somehow, this works.
Born in Morocco, Ghita left at 17 to study at École Polytechnique in Paris — France's most elite mathematics and engineering institution. She followed that with a master's in Applied Mathematics at UC Berkeley, where the idea for ZeroEntropy quietly took shape. Before ChatGPT went viral, she was already building conversational AI — and hitting the retrieval wall every time. That frustration became a company.
At 25, she is one of the very few female CEOs building deep AI infrastructure. Featured in TechCrunch and The AI Insider. Actively inspiring young women across Morocco and Africa to pursue engineering careers.
École Polytechnique · UC Berkeley · TechCrunchGrew up doing math and coding competitions. Dropped out of CMU to build startups. Became CTO at five different companies. Wrote low-level C, C++, Assembly and GPU code. Built stat-arb algorithms. Audited blockchains for bug bounties and hedge funds. At some point he built the AI at Myko.ai, Manifestapp and MagiBook.
The kind of engineer who treats a GPU as a musical instrument. Brought to ZeroEntropy the deep systems expertise that made zerank and zembed possible.
CMU · 5x CTO · Competitive CoderZeroEntropy's earliest backer was Entrepreneurs First, which pre-seeded the company before the YC application. Both founders had spent years inside AI projects — in healthcare, finance, B2B SaaS, and consumer apps — encountering the same retrieval failures again and again. Every team was reinventing the same broken pipeline. They decided to fix it once, for everyone.
Y Combinator's Winter 2025 batch took them in. The $4.2M seed round that followed was led by Initialized Capital, with participation from Transpose Platform, 22 Ventures, a16z Scout, and angels from OpenAI, Hugging Face, and Front.
They don't build everything. They build the part everyone else gets wrong.
A cross-encoder neural reranker trained with the proprietary zELO method — an Elo-scoring system originally used for chess ratings, now applied to query-document relevance. Beats Cohere rerank 3.5 and Jina rerank m0 in both speed and accuracy. About 12% faster on small payloads, 31% faster on large ones. Available via API, HuggingFace, AWS SageMaker, and Azure Marketplace.
A 4-billion parameter open-weight multilingual embedding model distilled directly from zerank-2 — meaning its relevance intuition is inherited from the reranker itself, not just binary labels. Supports 50+ languages. Compresses from 2,560 dimensions all the way down to 40. Reduces vector storage costs by up to 10x. Outperforms OpenAI, Cohere, Google, and Voyage on finance, healthcare, and legal benchmarks.
ZeroEntropy's full search engine: ingestion, preprocessing, hybrid retrieval, embedding, and reranking — all in one API. Handles negated queries ("articles NOT mentioning Elon Musk"), multi-hop queries, and fuzzy filtering. A Python SDK or interactive dashboard. No pipeline to stitch. No vector DB to babysit.
Run the entire ZeroEntropy stack inside your own cloud. No data leaves your VPC. SOC 2 Type II certified. HIPAA-ready. 99.99% SLA. White-glove onboarding and custom integrations. Available on AWS SageMaker and Azure Marketplace. Private offers for volume pricing and BAAs.
It's embarrassingly simple. That's the point.
NDCG@10 on MSMARCO — the closest public proxy to real RAG workloads. Higher = better.
Source: Agentset Leaderboard & ZeroEntropy public benchmarks. Scores approximate for competitor models.
Legal. Healthcare. Finance. Customer Support. Sales. The ones where wrong answers have consequences.
LegalBench-RAG — the first open-source legal retrieval benchmark — was built by ZeroEntropy. 6,800+ queries. 79M+ characters. Human-annotated spans. Contract understanding, case law retrieval, and regulatory search at scale.
Clinical documentation search, diagnostic support, medical literature retrieval. zembed-1 shows particularly strong domain performance on healthcare vocabulary and nuanced relevance ranking. HIPAA-ready on ze-onprem.
Earnings call analysis, research note retrieval, risk document search. zembed-1 outperforms every competitor on finance-domain benchmarks. Nicholas has built stat-arb algorithms and audited blockchains — they know this space.
Assembled — the support platform trusted by Stripe, Canva, Robinhood, and Notion — replaced their entire retrieval stack with ZeroEntropy. Result: better accuracy, lower latency, same scale. Zero rebuilding of pipelines.
zsearch is purpose-built for agentic workflows. Agents ask weird, multi-hop, negated, fuzzy questions. ZeroEntropy's system routes each query to the right retrieval strategy automatically. No hardcoded logic. Just accurate answers.
Users scan the top 10 results. Every misranked product is a lost sale. zerank's 60ms latency and class-leading NDCG@10 means better product discovery without the latency tax that kills conversions.
Small team. Big leaderboard presence.
Ghita and Nicholas, fresh from AI engineering roles across healthcare, finance, and SaaS, decide to fix retrieval once and for all. Pre-seeded by Entrepreneurs First.
ZeroEntropy joins YC's winter cohort. The combination of math-competition credentials and real production experience stands out in a crowded AI infrastructure field.
First public rerankers. zerank-1-small goes fully open-source under Apache 2.0. zerank-1 benchmarks show NDCG improvements of up to 5 points over Cohere, Voyage, and Salesforce's models. Outperforms Gemini Flash 2.0 as a reranker.
Led by Initialized Capital. YC, Transpose Platform, 22 Ventures, a16z Scout, and angels from OpenAI, HuggingFace, and Front join. TechCrunch covers the round: "Moroccan founder raises $4.2M for her YC-backed startup."
zerank-2 ships and immediately tops reranking leaderboards. Assembled — whose platform serves Stripe, Canva, Robinhood, and Notion — switches 100% of production retrieval reranking to ZeroEntropy.
The 4B open-weight multilingual embedding model ships and immediately tops the Agentset Leaderboard. Outperforms OpenAI, Google, Cohere, and Voyage on general retrieval and multilingual tasks.
ZeroEntropy's HuggingFace page says it plainly: "a team of mathematicians, physicists, and competitive programmers." This isn't branding. zerank-1 was initialized from Qwen3-4B and trained with a novel zELO pipeline derived from the Thurstone statistical model. zembed-1 was distilled from the reranker itself, inheriting calibrated relevance judgments that binary training labels cannot produce.
The hiring bar is explicit: they are actively recruiting a Head of Developer Experience — someone technical who can make complex ideas clear and enjoys being around builders. The model: deep technical, developer-first, no hand-waving.
Ghita uses her platform to speak openly about diversity in deep tech — particularly the near-absence of women in AI infrastructure. Her journey from Morocco to École Polytechnique to YC is part of ZeroEntropy's story, not a footnote to it.
If you are building AI agents, RAG pipelines, internal search, chatbots, or search bars at scale — and your retrieval layer is currently duct-tape and prayers — ZeroEntropy will save you months of engineering. Contact: founders@zeroentropy.dev
The $4.2M seed is deployed. The models are shipping and topping leaderboards. The next round will be oversubscribed. If you want in early on the infrastructure layer every AI product needs — reach out now. AWS, Azure, and HuggingFace are already distribution channels.
If you find zELO methodology interesting, if you've published on RAG or retrieval, if you've won a coding competition — they want to hear from you. The team is small. The problems are genuinely hard. The output is open-weight and published.
The name ZeroEntropy comes from information theory: high entropy = chaos, unpredictability, noise. Zero entropy = perfect knowledge of what a message contains. That's the philosophical claim they're making about search. Not "better search." Certain search. The zELO training method — their most important technical innovation — takes its name from the same Elo rating system used in chess, applied to document relevance scoring. The company is, essentially, a physics-meets-chess-meets-mathematics metaphor made into software.