Following the completion of this course, participants will:
This course is designed for:
3 Modules – 48 Videos – 17 Readings – 3 Assignments – 4 App Items – Certificate of Completion
This introductory program will take you on a thorough exploration of the principles of generative AI and how it may be used in practical situations. You will start the process by examining the application cases and tasks of generative AI and LLMs. Students will explore the revolutionary transformers architecture that changed the text generation industry. The lesson describes the project lifecycle for generative AI and covers the topic of producing text with transformers. It also covers the fundamental ideas of prompting and prompt engineering.
You will be given an overview of AWS labs and an in-depth overview of Lab 1, which focuses on summarizing dialogue as a generative AI use case, to provide you with hands-on experience. The module also covers the computational difficulties associated with training big language models, including information on effective multi-GPU compute techniques and scaling rules for models that are compute-optimal. In order to comprehend domain-specific training, you will also learn about pre-training for domain adaptation using case studies.
In the second module of this course, you will delve further into the basics of optimizing and assessing generative AI models. Students will learn about fine-tuning instructions, as well as single-task and multi-task instruction fine-tuning follow. You will discover how to analyze model performance in an efficient manner and the significance of benchmarks for determining the potential of generative AI.
This module places a lot of emphasis on parameter-efficient fine-tuning (PEFT). Several PEFT approaches will be covered, such as LoRA (Low-Rank Adaptation) and Soft Prompts, which allow huge language models to be efficiently fine-tuned without requiring a lot of processing power. A thorough tutorial of Lab 2, where you will refine a generative AI model for dialogue summary, emphasizes practical application will also be covered in this module.
You will examine the vital aspect of generative AI models' alignment with human values in the third module of this course, with a particular emphasis on reinforcement learning from human feedback (RLHF). The session starts by outlining the significance of matching models to human values. It then explores the several elements of reinforcement learning-based human-model fit (RLHF), such as getting human input, creating a reward model, and optimizing models using reinforcement learning. You will also discover how to successfully scale human feedback and the difficulties associated with reward hacking.
This module's detailed overview of Lab 3, where you use reinforcement learning to fine-tune the FLAN-T5 model to provide more favorable summaries, is a major highlight. Discussions on model optimizations for deployment, utilizing LLMs in applications, and connecting with external apps supplement this practical experience. Advanced topics covered in the curriculum include using program-aided language models (PAL), chain-of-thought approaches to assist LLMs in reasoning and planning, and the ReAct framework to integrate action and reasoning.
Learn the principles of generative AI and how to use it practically in real-world applications with the "Generative AI with Large Language Models (LLMs)" course. The goal of this course is to provide you with a thorough grasp of generative AI and how to use it to your advantage.
Developers can quickly create working prototypes and make wise decisions for their organizations if they have a strong basic grasp of LLMs and the best methods for training and deploying them. You will get a practical understanding in this training on how to make the most of this innovative new technology.
Since this is an intermediate-level course, you must have prior Python coding expertise. Additionally, you have to understand the fundamentals of machine learning, including loss functions, supervised and unsupervised learning, and data splitting into test, validation, and training sets.
Chris Fregly works at Amazon Web Services (AWS) in San Francisco, California, as a Principal Solutions Architect for Generative AI. He co-authored the O'Reilly books "Generative AI on AWS" and "Data Science on AWS." In addition, Chris is the creator of the international "Generative AI on AWS" Meetup series and a regular speaker at international AI and machine learning conferences, such as Nvidia GPU Technology Conference (GTC), O'Reilly AI, and Open Data Science Conference (ODSC).
Shelbee Eigenbrode works for Amazon Web Services (AWS) as a Principal Solutions Architect for Generative AI. She focuses on integrating her DevOps and ML skills to deploy and manage MLOps/FMOps/LLMOps workloads at scale. She has 23 years of experience in a variety of sectors, technologies, and positions, and she holds six AWS certifications. Shelbee is enthusiastic about constant innovation and using data to drive business outcomes. She has over 30 patents in a variety of technological disciplines. She is a co-founder of Women in Big Data in Denver and a published author.
At Amazon Web Services (AWS), Antje Barth is a Principal Developer Advocate for Generative AI. She co-wrote "Generative AI on AWS" and "Data Science on AWS." Speaking often at AI/ML conferences, events, and meetups throughout the world is Antje. Prior to joining Cisco and MapR, she worked in technical evangelism and solutions engineering, specializing in big data, AI applications, and data center technologies. Women in Big Data's Düsseldorf branch was co-founded by Antje.
Mike Chambers works for Amazon Web Services (AWS) as a Developer Advocate for Generative AI. Mike has worked as a solutions and security architect for more than 20 years in the ICT sector, having worked with major corporations, governments, and his own prosperous start-up. With his distinct teaching style, Mike has instructed and amused more than a quarter of a million students worldwide as a trainer, specializing in cloud, serverless, and machine learning technologies.