Italian, adopted by France not long ago, I am a constant learner, dedicated to computer science and discovery—whether uncovering solutions or gaining insights.
Semantic search is often hailed as a game-changer, promising to solve challenges like relevance, complex sentence analysis, and synonym detection with just a few embeddings and a machine learning model. The demos look impressive—but what happens when you're dealing with more than a billion embeddings?
In this talk, we move past the hype to explore the real-world complexities of managing large-scale vector databases, focusing on Elasticsearch and OpenSearch. Through practical, hands-on examples, we’ll share proven strategies to ensure scalability, maintain high performance, and optimize costs. Whether you're already managing a billion-vector database or preparing for large-scale deployment, this session will equip you with the knowledge and tools to tackle real-world challenges effectively.
Search technology has evolved from matching words to understanding meaning. In this talk, we’ll unravel how this transformation happened, from the early days of sparse vector models like TF-IDF and BM25, which relied purely on keyword overlap, to today’s dense vector embeddings powered by neural networks and transformers. We’ll explore how semantic search captures context, relationships, and intent, allowing machines to “understand” language rather than just count words.
Using intuitive examples and visual demonstrations, we’ll demystify how embeddings represent meaning in high-dimensional space, how similarity is computed, and why this shift is revolutionizing information retrieval, recommendation, and question answering.
Attendees will leave with a clear conceptual map of how sparse and dense approaches differ, and how they can work together to build smarter, more intuitive search systems.
Searching for speaker images...
