Schedule

Join us in Charlottesville, VA, for Haystack US 2026 – an immersive experience featuring workshops and two days of conference talks tailored for search and AI professionals.

Schedule

Filter by:

Track

No tracks available

Session Type

Wednesday, 6. May

Wednesday
Main Stage
09:00
09:30

Welcome

Main Stage
Talk
09:00 - 09:30 (30m)
Welcome
09:30
10:15

AI Governance: Crafting Your Own AI Experiences

Main Stage
Talk
09:30 - 10:15 (45m)
In a world where AI shapes what we see, think, and do, true enginering lies not in using tools—but in designing them.
10:20
11:05

Learning to Understand: A Missing Stage of Modern Retrieval

Main Stage
Talk
10:20 - 11:05 (45m)
We introduce “Learning to Understand” as a corollary to the well known “Learning to Rank” process. By using evals to learn domain-specific query interpretation and rewriting rules and combining with semantic statistics from your index, it’s possible to significantly improve search quality beyond typical BM25, vector, and hybrid search techniques.
11:15
12:00

From 0 - Production with BBQ at GitHub

Main Stage
Talk
11:15 - 12:00 (45m)
Rolling out semantic search is easy right? Just turn on some vectors and bim bam boom you have vector search... Right? Turns out when you're GitHub sized it's not quite that easy. We'll walk through the process we took, the lessons we've learned, and how you can build a plan to deploy vector search easier.
14:20
15:05

Why your B2B search engine doesn’t understand your users

Main Stage
Talk
14:20 - 15:05 (45m)
This talk uses a real-world B2B search case to show how a decision-based tree helps quickly diagnose why search fails and how to improve relevance without rebuilding the system.
15:20
16:05

Search rankers coded by agents

Main Stage
Talk
15:20 - 16:05 (45m)
Could AI code generation replace Learning to Rank? AI coding tools can generate rankers, but only up to a point. What techniques matter when building an agent coded ranker? And where do traditional search techniques still work?
16:10
17:25

Lightning Talks

Main Stage
Talk
16:10 - 17:25 (1h 15m)
Lightning Talks

Thursday, 7. May

Thursday
Main Stage
09:15
10:00

Agentic Tuning: Search Relevance on Autopilot

Main Stage
Talk
09:15 - 10:00 (45m)
Search relevance tuning is notoriously difficult, often requiring a deep understanding of Lucene scoring, complex query DSLs, and iterative manual testing. This session introduces Agentic Relevance Tuning—a framework that leverages LLM-based agents to automate the full search lifecycle making search tuning faster, more accurate, and accessible.
10:05
10:50

Evolution of Relevance Engineering to Context Engineering

Main Stage
Talk
10:05 - 10:50 (45m)
As search powers RAG and agentic systems, relevance goals shift from ranking documents to assembling effective context. This talk explores how traditional lexical, semantic, and hybrid relevance changes when feeding LLMs, with lessons on chunking and snippet extraction, diversification, evaluation, and more.
11:00
11:45

LLMs as Rerankers: A Case Study on Hybrid Email Search

Main Stage
Talk
11:00 - 11:45 (45m)
Purpose-built rerankers are faster and cheaper, but are they better? We argue LLM rerankers win on what matters most in production: instruction-following and iteration speed, with more-than-acceptable tradeoffs on cost and latency. Our discussion is backed by a case study from Superhuman's production hybrid email search system.
13:15
14:00

Do we still need search engines?

Main Stage
Talk
13:15 - 14:00 (45m)
Search has a new User Interface! All search will be Agentic/RAG and delivered through a chat interface! The days of the monolithic search engine are over!
14:05
14:50

Managing Search Teams: Field Stories & Practical Takeaways

Main Stage
Talk
14:05 - 14:50 (45m)
Even though search teams are structured differently across the industry, they share common challenges like balancing learning with delivery and nurturing a culture built for continuous iteration. This talk distills a decade of organizational lessons from building Yelp’s AI-powered search into repeatable patterns for any team facing similar hurdles.
15:05
15:50

Adaptive Relevance with Agentic Search

Main Stage
Talk
15:05 - 15:50 (45m)
Traditional search pipelines rely heavily on static query parsing and after-the-fact relevance analysis. In this session, we present a new paradigm: using LangGraph with OpenSearch to create an agentic-based system that can tune hybrid search in real time.
16:00
16:45

When BM25 Scores Disagree: A Corpus-Independent Alternative

Main Stage
Talk
16:00 - 16:45 (45m)
In distributed search, BM25 returns different results across nodes because IDF and average document length vary with each node's corpus state. StableTfl replaces these with a term-length rarity heuristic, eliminating all corpus dependency. On 22 BEIR datasets, it retains ~90% of BM25's NDCG@10 while guaranteeing identical rankings across nodes.
16:45
17:00

Closing

Main Stage
Talk
16:45 - 17:00 (15m)
Closing