• 16: Infini-Attention: Google's Solution for Infinite Memory in LLMs

  • May 22 2024
  • Duración: 23 m
  • Podcast

16: Infini-Attention: Google's Solution for Infinite Memory in LLMs  Por  arte de portada

16: Infini-Attention: Google's Solution for Infinite Memory in LLMs

  • Resumen

  • In this episode of the AI Paper Club Podcast, hosts Rafael Herrera and Sonia Marques welcome Leticia Fernandes, a Deeper Insights Senior Data Scientist and Generative AI Ambassador. Together, they explore the groundbreaking "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" paper from Google. This paper addresses the challenge of fitting infinite context into large language models, introducing the Infini-attention method. The trio discusses how this approach works, including how it uses linear attention and employs compressive memory to store key-value pairs, enabling models to handle extensive contexts.

    We also extend a special thank you to the research team Google for developing this month’s paper. If you are interested in reading the paper for yourself, please check this link: https://arxiv.org/pdf/2404.07143.pdf

    For more information on all things artificial intelligence, machine learning, and engineering for your business, please visit www.deeperinsights.com or reach out to us at thepaperclub@deeperinsights.com.


    Más Menos
activate_primeday_promo_in_buybox_DT

Lo que los oyentes dicen sobre 16: Infini-Attention: Google's Solution for Infinite Memory in LLMs

Calificaciones medias de los clientes

Reseñas - Selecciona las pestañas a continuación para cambiar el origen de las reseñas.