• Abstracts: NeurIPS 2024 with Weizhu Chen 

  • Dec 6 2024
  • Length: 8 mins
  • Podcast

Abstracts: NeurIPS 2024 with Weizhu Chen 

  • Summary

  • Next-token prediction trains a language model on all tokens in a sequence. VP Weizhu Chen discusses his team’s 2024 NeurIPS paper on how distinguishing between useful and “noisy” tokens in pretraining can improve token efficiency and model performance.

    Read the paper

    Get the code

    Show More Show Less
activate_Holiday_promo_in_buybox_DT_T2

What listeners say about Abstracts: NeurIPS 2024 with Weizhu Chen 

Average customer ratings

Reviews - Please select the tabs below to change the source of reviews.