Efficiently Scaling and Deploying LLMs // Hanlin Tang // LLM's in Production Conference

Поділитися
Вставка
  • Опубліковано 21 сер 2024
  • // Abstract
    Hanlin discusses the evolution of Large Language Models and the importance of efficient scaling and deployment. He emphasizes the benefits of a decentralized approach of many small specialized models over one giant AGI model controlled by a few companies. Hanlin explains the advantages of companies training their own custom models, such as data privacy concerns, and provides insights into when it is appropriate to build your own models and the available tooling for training and deployment.
    // Bio
    Hanlin is the CTO & Co-founder of MosaicML, an ML infrastructure startup that enables enterprises to easily train large-scale AI models in their secure environments. Hanlin was previously the Director of the Intel AI Lab, responsible for the research and deployment of deep learning models. He joined Intel from its acquisition of Nervana Systems. Hanlin has a Ph.D. from Harvard University and has published in leading journals and conferences such as NeurIPS, ICLR, ICML, Neuron, and PNAS.

КОМЕНТАРІ • 3

  • @reyruben619
    @reyruben619 Рік тому +2

    Love it! Would love to build a maintenance chatbot with one llm. Many opportunities 💪

  • @MLOps
    @MLOps  Рік тому +1

    We are having another *LLM in production* event with some pretty amazing speakers. sign up here home.mlops.community/home/events/llm-in-prod-part-ii-2023-06-20

  • @yamnel
    @yamnel 4 місяці тому

    Would have loved information about actually Scaling and Deploying LLMs in production...