mirror of
https://github.com/karpathy/nanoGPT.git
synced 2026-04-17 14:19:18 +02:00
Update README to mention nanochat and deprecation
Added a note about the deprecation of nanoGPT and the introduction of nanochat.
This commit is contained in:
@@ -3,6 +3,13 @@
|
||||
|
||||

|
||||
|
||||
|
||||
---
|
||||
|
||||
**Update Nov 2025** nanoGPT has a new and improved cousin called [nanochat](https://github.com/karpathy/nanochat). It is very likely you meant to use/find nanochat instead. nanoGPT (this repo) is now very old and deprecated but I will leave it up for posterity.
|
||||
|
||||
---
|
||||
|
||||
The simplest, fastest repository for training/finetuning medium-sized GPTs. It is a rewrite of [minGPT](https://github.com/karpathy/minGPT) that prioritizes teeth over education. Still under active development, but currently the file `train.py` reproduces GPT-2 (124M) on OpenWebText, running on a single 8XA100 40GB node in about 4 days of training. The code itself is plain and readable: `train.py` is a ~300-line boilerplate training loop and `model.py` a ~300-line GPT model definition, which can optionally load the GPT-2 weights from OpenAI. That's it.
|
||||
|
||||

|
||||
|
||||
Reference in New Issue
Block a user