Suggested Readings! 

  • Hamiltonian Neural Networks

    Hamiltonian Neural Networks

    Drawing inspiration from Hamiltonian mechanics, a branch of physics concerned with conservation laws and invariances, we define Hamiltonian Neural Networks, or HNNs. By construction, these models learn conservation laws from data. We will show that they have some major advantages over regular neural networks on a variety of physics problems.

    45 min read
  • Introducing DoRA, a High-Performing Alternative to LoRA for Fine-Tuning

    Introducing DoRA, a High-Performing Alternative to LoRA for Fine-Tuning

    Full fine-tuning (FT) is commonly employed to tailor general pretrained models for specific downstream tasks. To reduce the training cost, parameter-efficient fine-tuning (PEFT) methods have been introduced to fine-tune pretrained models with a minimal number of parameters. Among these, Low-Rank Adaptation (LoRA) and its variants have gained considerable popularity because they avoid additional inference costs.

    30 min read
  • Alice's Adventures in a Differentiable Wonderland

    Alice's Adventures in a Differentiable Wonderland

    Neural networks surround us, in the form of large language models, speech transcription systems, molecular discovery algorithms, robotics, and much more. Stripped of anything else, neural networks are compositions of differentiable primitives, and studying them means learning how to program and how to interact with these models, a particular example of what is called differentiable programming. This primer is an introduction to this fascinating field imagined for someone, like Alice, who has just ventured into this strange differentiable wonderland.

    Book - 200 pages
  • Contrastive Language-Image Pre-training

    Contrastive Language-Image Pre-training

    CLIP (Contrastive Language-Image Pre-training) builds on a large body of work on zero-shot transfer, natural language supervision, and multimodal learning. The idea of zero-data learning dates back over a decade8 but until recently was mostly studied in computer vision as a way of generalizing to unseen object categories.

    20 min read
  • Learning to play Minecraft with Video PreTraining

    Learning to play Minecraft with Video PreTraining

    We trained a neural network to play Minecraft by Video PreTraining (VPT) on a massive unlabeled video dataset of human Minecraft play, while using only a small amount of labeled contractor data.

    20 min read

All articles