Self-attention Bilibili Li Hongyi’s explanation video: https://www.bilibili.com/video/BV1Xp4y1b7ih
KV Cache introduction: https://zhuanlan.zhihu.com/p/630832593
This is an incomplete implementation of GPT-2.
A quick breakdown of each of the files:
encoder.py
contains the code for OpenAI's BPE Tokenizer.utils.py
contains the code to download and load the GPT-2 model weights, tokenizer, and hyper-parameters.NSL-gpt2.py
contains the actual GPT model and generation code which we can run as a python script, but it is an incomplete version. Believe that you can successfully complete it😎👍.
pip install -r requirements.txt
The first run requires downloading the model, which is slow, please be patient.
python NSL-gpt2.py \
"Alan Turing theorized that computers would one day become" \
--n_tokens_to_generate 40
Which generates
the most powerful machines on the planet.
The computer is a machine that can perform complex calculations, and it can perform these calculations in a way that is very similar to the human brain.
We used 124M model in this test. You can also control the number of tokens to generate, the model size (one of ["124M", "355M", "774M", "1558M"]
), and the directory to save the models:
python NSL-gpt2.py \
"Alan Turing theorized that computers would one day become" \
--n_tokens_to_generate 40 \
--model_size "124M" \
--models_dir "models"
When you write the greedy_speculative_decoding function, you need to load both the 124M and 1558M models at the same time, and be careful to modify the parameters when loading the models.
When we use the 1558M model for autoregressive inference, the returned results are as follows: (When using greedy sampling, this result is unique)
so powerful that they would be able to think like humans.
In the 1950s, he proposed a way to build a computer that could think like a human. He called it the "T