Coding LLaMA-2 from scratch in PyTorch - Part 1

Описание к видео Coding LLaMA-2 from scratch in PyTorch - Part 1

In this video series, you will learn how to train and fine-tune Llama 2 model from scrach.

The goal is to code LLaMA 2 from scratch in PyTorch to create models with sizes 100M, 250M and 500M params. In this first video, you'll learn about transformer archictecture in detail and implement a basic model with 100M params using PyTorch.

This is a step-by-step guide to Llama 2 model implementation based on the research paper.

To follow along you can use this colab notebook:
https://colab.research.google.com/dri...

Комментарии

Информация по комментариям в разработке