English | MP4 | AVC 1920×1080 | AAC 44KHz 2ch | 15h 3m | 2.69 GB
Apply cutting-edge machine learning techniques—from crowdsourced relevance and knowledge graph learning, to Large Language Models (LLMs)—to enhance the accuracy and relevance of your search results.
Delivering effective search is one of the biggest challenges you can face as an engineer. AI-Powered Search is an in-depth guide to building intelligent search systems you can be proud of. It covers the critical tools you need to automate ongoing relevance improvements within your search applications.
Inside you’ll learn modern, data-science-driven search techniques like:
- Semantic search using dense vector embeddings from foundation models
- Retrieval augmented generation (RAG)
- Question answering and summarization combining search and LLMs
- Fine-tuning transformer-based LLMs
- Personalized search based on user signals and vector embeddings
- Collecting user behavioral signals and building signals boosting models
- Semantic knowledge graphs for domain-specific learning
- Semantic query parsing, query-sense disambiguation, and query intent classification
- Implementing machine-learned ranking models (Learning to Rank)
- Building click models to automate machine-learned ranking
- Generative search, hybrid search, multimodal search, and the search frontier
AI-Powered Search will help you build the kind of highly intelligent search applications demanded by modern users. Whether you’re enhancing your existing search engine or building from scratch, you’ll learn how to deliver an AI-powered service that can continuously learn from every content update, user interaction, and the hidden semantic relationships in your content. You’ll learn both how to enhance your AI systems with search and how to integrate large language models (LLMs) and other foundation models to massively accelerate the capabilities of your search technology.
Modern search is more than keyword matching. Much, much more. Search that learns from user interactions, interprets intent, and takes advantage of AI tools like large language models (LLMs) can deliver highly targeted and relevant results. This book shows you how to up your search game using state-of-the-art AI algorithms, techniques, and tools.
AI-Powered Search teaches you to create a search that understands natural language and improves automatically the more it is used. As you work through dozens of interesting and relevant examples, you’ll learn powerful AI-based techniques like semantic search on embeddings, question answering powered by LLMs, real-time personalization, and Retrieval Augmented Generation (RAG).
What’s inside
- Sparse lexical and embedding-based semantic search
- Question answering, RAG, and summarization using LLMs
- Personalized search and signals boosting models
- Learning to Rank, multimodal, and hybrid search
Table of Contents
1 Part 1. Modern search relevance
2 Chapter 1. Introducing AI-powered search
3 Chapter 1. Understanding user intent
4 Chapter 1. How does AI-powered search work
5 Chapter 1. Summary
6 Chapter 2. Working with natural language
7 Chapter 2. The structure of natural language
8 Chapter 2. Distributional semantics and embeddings
9 Chapter 2. Modeling domain-specific knowledge
10 Chapter 2. Challenges in natural language understanding for search
11 Chapter 2. Content + signals The fuel powering AI-powered search
12 Chapter 2. Summary
13 Chapter 3. Ranking and content-based relevance
14 Chapter 3. Controlling the relevance calculation
15 Chapter 3. Implementing user and domain-specific relevance ranking
16 Chapter 3. Summary
17 Chapter 4. Crowdsourced relevance
18 Chapter 4. Introducing reflected intelligence
19 Chapter 4. Summary
20 Part 2. Learning domain-specific intent
21 Chapter 5. Knowledge graph learning
22 Chapter 5. Using our search engine as a knowledge graph
23 Chapter 5. Automatically extracting knowledge graphs from content
24 Chapter 5. Learning intent by traversing semantic knowledge graphs
25 Chapter 5. Using knowledge graphs for semantic search
26 Chapter 5. Summary
27 Chapter 6. Using context to learn domain-specific language
28 Chapter 6. Query-sense disambiguation
29 Chapter 6. Learning related phrases from query signals
30 Chapter 6. Phrase detection from user signals
31 Chapter 6. Misspellings and alternative representations
32 Chapter 6. Pulling it all together
33 Chapter 6. Summary
34 Chapter 7. Interpreting query intent through semantic search
35 Chapter 7. Indexing and searching on a local reviews dataset
36 Chapter 7. An end-to-end semantic search example
37 Chapter 7. Query interpretation pipelines
38 Chapter 7. Summary
39 Part 3. Reflected intelligence
40 Chapter 8. Signals-boosting models
41 Chapter 8. Normalizing signals
42 Chapter 8. Fighting signal spam
43 Chapter 8. Combining multiple signal types
44 Chapter 8. Time decays and short-lived signals
45 Chapter 8. Index-time vs. query-time boosting Balancing scale vs. flexibility
46 Chapter 8. Summary
47 Chapter 9. Personalized search
48 Chapter 9. Recommendation algorithm approaches
49 Chapter 9. Implementing collaborative filtering
50 Chapter 9. Personalizing search using content-based embeddings
51 Chapter 9. Challenges with personalizing search results
52 Chapter 9. Summary
53 Chapter 10. Learning to rank for generalizable search relevance
54 Chapter 10. Step 1 A judgment list, starting with the training data
55 Chapter 10. Step 2 Feature logging and engineering
56 Chapter 10. Step 3 Transforming LTR to a traditional machine learning problem
57 Chapter 10. Step 4 Training (and testing!) the model
58 Chapter 10. Steps 5 and 6 Upload a model and search
59 Chapter 10. Rinse and repeat
60 Chapter 10. Summary
61 Chapter 11. Automating learning to rank with click models
62 Chapter 11. Overcoming position bias
63 Chapter 11. Handling confidence bias Not upending your model due to a few lucky clicks
64 Chapter 11. Exploring your training data in an LTR system
65 Chapter 11. Summary
66 Chapter 12. Overcoming ranking bias through active learning
67 Chapter 12. AB testing a new model
68 Chapter 12. Overcoming presentation bias Knowing when to explore vs. exploit
69 Chapter 12. Exploit, explore, gather, rinse, repeat A robust automated LTR loop
70 Chapter 12. Summary
71 Part 4. The search frontier
72 Chapter 13. Semantic search with dense vectors
73 Chapter 13. Search using dense vectors
74 Chapter 13. Getting text embeddings by using a Transformer encoder
75 Chapter 13. Applying Transformers to search
76 Chapter 13. Natural language autocomplete
77 Chapter 13. Semantic search with LLM embeddings
78 Chapter 13. Quantization and representation learning for more efficient vector search
79 Chapter 13. Cross-encoders vs. bi-encoders
80 Chapter 13. Summary
81 Chapter 14. Question answering with a fine-tuned large language model
82 Chapter 14. Constructing a question-answering training dataset
83 Chapter 14. Fine-tuning the question-answering model
84 Chapter 14. Building the reader with the new fine-tuned model
85 Chapter 14. Incorporating the retriever Using the question-answering model with the search engine
86 Chapter 14. Summary
87 Chapter 15. Foundation models and emerging search paradigms
88 Chapter 15. Generative search
89 Chapter 15. Multimodal search
90 Chapter 15. Other emerging AI-powered search paradigms
91 Chapter 15. Hybrid search
92 Chapter 15. Convergence of contextual technologies
93 Chapter 15. All the above, please!
94 Chapter 15. Summary
95 Appendix A. Running the code examples
96 Appendix A. Pulling the source code
97 Appendix A. Building and running the code
98 Appendix A. Working with Jupyter
99 Appendix A. Working with Docker
100 Appendix B. Supported search engines and vector databases
101 Appendix B. Swapping out the engine
102 Appendix B. The engine and collection abstractions
103 Appendix B. Adding support for additional engines
Resolve the captcha to access the links!