Build Long-Context AI Apps with Jamba
Instructors: Chen Wang, Chen Almagor
- Beginner
- 1 Hour 3 Minutes
- 9 Video Lessons
- 5 Code Examples
- Instructors: Chen Wang, Chen Almagor
What you'll learn
Learn how the Jamba model integrates transformers and the Mamba architecture to efficiently process long contexts while maintaining quality.
Understand the training process for long context models, and the metrics used to evaluate their performance.
Gain hands-on experience applying Jamba to tasks such as processing large documents, tool-calling, and building large context RAG apps.
About this course
Learn to use the Jamba model, a hybrid transformer-Mamba architecture trained to handle long contexts, in this new course, Build Long-Context AI Apps with Jamba, built in partnership with AI21 Labs and taught by Chen Wang and Chen Almagor.
The transformer architecture is the foundation of most large language models but it is computationally expensive when handling very long input contexts.
There’s an alternative to transformers called Mamba. Mamba is a selective state space model that can process very long contexts with a much lower computational cost. However, researchers found that the pure Mamba architecture underperforms in understanding the context, and the quality of the output can be lower even in tasks as simple as repeating the input in the output of the model.
To overcome these challenges, AI21 Labs developed the Jamba model, which combines Mamba’s computational efficiency with the transformer’s attention mechanism to help with the output quality.
In this course, you’ll learn about the Jamba architecture, how it works, and how it is trained. You’ll also learn how to prompt Jamba and use it to process long documents and build long-context RAG apps.
In detail, you’ll:
- Learn key information about Jamba’s large context model and its unique, hybrid architecture that allows for high-quality outputs, high throughput, and low memory usage.
- Learn how Jamba combines transformers with Mamba, a selective state-space model to achieve high performance and quality.
- Use the AI21SDK, especially its document parameter, with an example of prompting a large over 200k-token annual report.
- Use Jamba for tool-calling with hands-on examples from calling simple arithmetic functions to a function that returns SEC 10-Q company quarterly report.
- Learn how the training for long context is done, and the metrics used for evaluating the performance.
- Create a RAG app using the AI21 Conversational RAG tool and build your own RAG pipeline that uses the Jamba model with LangChain.
Start building AI apps that can handle context as long as all of your unread emails from the last 20 years!
Who should join?
Anyone who has basic Python knowledge and wants to learn more about how the Jamba model works and how it is used for building long-context AI apps.
Course Outline
9 Lessons・5 Code ExamplesIntroduction
Video・3 mins
Overview
Video・5 mins
Transformer-Mamba Hybrid LLM Architecture
Video・14 mins
Jamba Prompting and Documents
Video with code examples・8 mins
Tool Calling
Video with code examples・6 mins
Expand the Context Window Size
Video・13 mins
Long Context Prompting
Video with code examples・3 mins
Conversational RAG
Video with code examples・8 mins
Conclusion
Video・1 min
Appendix – Tips and Help
Code examples・1 min
Instructors
Build Long-Context AI Apps with Jamba
- Beginner
- 1 Hour 3 Minutes
- 9 Video Lessons
- 5 Code Examples
- Instructors: Chen Wang, Chen Almagor
Course access is free for a limited time during the DeepLearning.AI learning platform beta!
Want to learn more about Generative AI?
Keep learning with updates on curated AI news, courses, and events, as well as Andrew’s thoughts from DeepLearning.AI!