ReaLLMASIC / nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs.
MIT License
24 stars 18 forks source link

Add pos emb finetuning option #263

Closed gkielian closed 2 months ago

gkielian commented 2 months ago

This will allow not setting the abs pos embedding table if one wants to instead finetune Rotary, functional embeddings, or other alternatives which allow for better context length extensions.

To be merged in after PR #262