The first content gpt2.py
and inference.py
together is a test project that follows Andrej Karpathy's GPT2 repreduction video lesson with some changes.
It uses a small portion of fineweb-edu dataset 600M tokens (100M validation)
with RTX 4090 Laptop GPU (16GB).
The Transformers
folder contains different implementation of GPT, a fine-tuning and a RAG example.
This file get some of the codes from Transformers for Natural Language Processing and Computer Vision from Denis Rothman
and Natural Language Processing with Transformers from Lewis Tunstall, Leandro von Werra, Thomas Wolf
.
Main purpose for this project is to make a series of internal company sessions about LLMs and their implementation details.
- Python==3.10.14
- Pytorch==2.3.1+cu121
To set up and running these projects:
- Create a conda environment.
- Install the requirements.
conda create --name <ENV_NAME> python=3.10.14 --file requirements.txt
👤 Eyüp Sercan UYGUR
Srjnnnn(Twitter)
eyupsercanuygur.com
Contributions, issues and feature requests are welcome!
Give a ⭐️ if you like this project!
For issues check.
This project is MIT licensed.