Baptiste Roziere

Baptiste Roziere

Research Scientist at Meta

    Baptiste is a research scientist at Meta AI in Paris working in the code generation team. He contributed to Llama and led Code Llama. At Meta, Baptiste conducted research on unsupervised translation of programming languages and model pre-training for code. His work was featured in dozens of news articles in more than ten languages. He also started a collaboration between the Fundamental AI Research department and production teams putting code models in production. Prior to Meta, Baptiste worked as an applied scientist in advertising at Amazon.

    All Sessions by Baptiste Roziere

    Day 1 04/23/2024
    12:10 pm - 12:40 pm

    CodeLlama: Open Foundation Models for Code

    <span class="etn-schedule-location"> <span class="firstfocus">Large Language Models</span>

    In this session, we will present the methods used to train Code Llama, the performance we obtained, and show how you could use Code Llama in practice for many software development use cases. Code Llama is a family of open large language models for code based on Llama 2 providing state-of-the-art performance among open models, infilling capabilities, support for large input contexts, and zero-shot instruction following ability for programming tasks. We provide multiple flavors to cover a wide range of applications: foundation models (Code Llama), Python specializations (Code Llama - Python), and instruction-following models (Code Llama - Instruct) with 7B, 13B, 34B, and now 70B parameters each. Code Llama reaches state-of-the-art performance among open models on several code benchmarks. Notably, Code Llama - Python 7B outperforms Llama 2 70B on HumanEval and MBPP, and all our models outperform every other open model on MultiPL-E. Code Llama was released under a permissive license that allows for both research and commercial use.

    Day 1 04/23/2024
    12:10 pm - 12:40 pm

    CodeLlama: Open Foundation Models for Code

    <span class="etn-schedule-location"> <span class="firstfocus">LLMs</span> </span>

    In this session, we will present the methods used to train Code Llama, the performance we obtained, and show how you could use Code Llama in practice for many software development use cases. Code Llama is a family of open large language models for code based on Llama 2 providing state-of-the-art performance among open models, infilling capabilities, support for large input contexts, and zero-shot instruction following ability for programming tasks. We provide multiple flavors to cover a wide range of applications: foundation models (Code Llama), Python specializations (Code Llama - Python), and instruction-following models (Code Llama - Instruct) with 7B, 13B, 34B, and now 70B parameters each. Code Llama reaches state-of-the-art performance among open models on several code benchmarks. Notably, Code Llama - Python 7B outperforms Llama 2 70B on HumanEval and MBPP, and all our models outperform every other open model on MultiPL-E. Code Llama was released under a permissive license that allows for both research and commercial use.

    Open Data Science

     

     

     

    Open Data Science
    One Broadway
    Cambridge, MA 02142
    info@odsc.com

    Privacy Settings
    We use cookies to enhance your experience while using our website. If you are using our Services via a browser you can restrict, block or remove cookies through your web browser settings. We also use content and scripts from third parties that may use tracking technologies. You can selectively provide your consent below to allow such third party embeds. For complete information about the cookies we use, data we collect and how we process them, please check our Privacy Policy
    Youtube
    Consent to display content from - Youtube
    Vimeo
    Consent to display content from - Vimeo
    Google Maps
    Consent to display content from - Google