Everything About Large Language Models: Pre-training, Fine-tuning, RLHF & State of the Art

Abstract: 

Generative Large Language Models like GPT4 have revolutionized the entire tech ecosystem. But what makes them so powerful? What are the secret components which make them generalize to a variety of tasks? In this talk, I will present how these foundation models are trained. What are the steps and core-components behind these LLMs? I will also cover how smaller, domain-specific models can outperform general purpose foundation models like ChatGPT on target use cases

Bio: 

Best known for developing state-of-the-art AI products such as the world’s first fully autonomous Conversational AI technology, the Alexa Prize (ChatGPT-like voice experience for Alexa users 5 years before ChatGPT), and Truth Checker AI, the first and currently only model to detect hallucinations generated by language models such as GPT-4.

Open Data Science

 

 

 

Open Data Science
One Broadway
Cambridge, MA 02142
info@odsc.com

Privacy Settings
We use cookies to enhance your experience while using our website. If you are using our Services via a browser you can restrict, block or remove cookies through your web browser settings. We also use content and scripts from third parties that may use tracking technologies. You can selectively provide your consent below to allow such third party embeds. For complete information about the cookies we use, data we collect and how we process them, please check our Privacy Policy
Youtube
Consent to display content from - Youtube
Vimeo
Consent to display content from - Vimeo
Google Maps
Consent to display content from - Google