| ## Converting LitGPT weights to Hugging Face Transformers |
|
|
| LitGPT weights need to be converted to a format that Hugging Face understands with a [conversion script](../litgpt/scripts/convert_lit_checkpoint.py) before our scripts can run. |
|
|
| We provide a helpful command to convert models LitGPT models back to their equivalent Hugging Face Transformers format: |
|
|
| ```bash |
| litgpt convert_from_litgpt checkpoint_dir converted_dir |
| ``` |
|
|
| These paths are just placeholders, you will need to customize them based on which finetuning or pretraining command you ran and its configuration. |
|
|
| ### Loading converted LitGPT checkpoints into transformers |
|
|
|
|
| For example, |
|
|
| ```bash |
| cp checkpoints/repo_id/config.json converted/config.json |
| ``` |
|
|
| Then, you can load the checkpoint file in a Python session as follows: |
|
|
| ```python |
| import torch |
| from transformers import AutoModel |
| |
| |
| state_dict = torch.load("output_dir/model.pth") |
| model = AutoModel.from_pretrained( |
| "output_dir/", local_files_only=True, state_dict=state_dict |
| ) |
| ``` |
|
|
| Alternatively, you can also load the model without copying the `config.json` file as follows: |
|
|
| ```python |
| model = AutoModel.from_pretrained("online_repo_id", state_dict=state_dict) |
| ``` |
|
|
|
|
|
|
| ### Merging LoRA weights |
|
|
| Please note that if you want to convert a model that has been finetuned using an adapter like LoRA, these weights should be [merged](../litgpt/scripts/merge_lora.py) to the checkpoint prior to converting. |
|
|
| ```sh |
| litgpt merge_lora path/to/lora/checkpoint_dir |
| ``` |
|
|
| <br> |
| <br> |
|
|
| # A finetuning and conversion tutorial |
|
|
| This section contains a reproducible example for finetuning a LitGPT model and converting it back into a HF `transformer` model. |
|
|
| 1. Download a model of interest: |
|
|
| For convenience, we first specify an environment variable (optional) to avoid copy and pasting the whole path: |
|
|
| ```bash |
| export repo_id=TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T |
| ``` |
|
|
| Instead of using TinyLlama, you can replace the `repo_id` target with any other model repository |
| specifier that is currently supported by LitGPT. You can get a list of supported repository specifier |
| by running `litgpt/scripts/download.py` without any additional arguments. |
|
|
| Then, we download the model we specified via `$repo_id` above: |
|
|
| ```bash |
| litgpt download $repo_id |
| ``` |
|
|
| 2. Finetune the model: |
|
|
|
|
| ```bash |
| export finetuned_dir=out/lit-finetuned-model |
| |
| litgpt finetune_lora $repo_id \ |
| --out_dir $finetuned_dir \ |
| --train.epochs 1 \ |
| --data Alpaca |
| ``` |
|
|
| 3. Merge LoRA weights: |
|
|
| Note that this step only applies if the model was finetuned with `lora.py` above and not when `full.py` was used for finetuning. |
|
|
| ```bash |
| litgpt merge_lora $finetuned_dir/final |
| ``` |
|
|
|
|
| 4. Convert the finetuning model back into a HF format: |
|
|
| ```bash |
| litgpt convert_from_litgpt $finetuned_dir/final/ out/hf-tinyllama/converted |
| ``` |
|
|
|
|
| 5. Load the model into a `transformers` model: |
|
|
| ```python |
| import torch |
| from transformers import AutoModel |
| |
| state_dict = torch.load('out/hf-tinyllama/converted/model.pth') |
| model = AutoModel.from_pretrained("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T", state_dict=state_dict) |
| ``` |
|
|
| |
| ## Using the LM Evaluation Harness |
|
|
| To evaluate LitGPT models, use the integrated evaluation utilities based on Eleuther AI's LM Evaluation Harness. For more information, please see the [evaluation](evaluation.md) documentation. |
|
|
| Alternatively, if you wish to use converted LitGPT models with the LM Evaluation Harness from [Eleuther AI's GitHub repository](https://github.com/EleutherAI/lm-evaluation-harness), you can use the following steps. |
|
|
| 1. Follow the instructions above to load the model into a Hugging Face transformers model. |
|
|
| 2. Create a `model.safetensor` file: |
|
|
| ```python |
| model.save_pretrained("out/hf-tinyllama/converted/") |
| ``` |
|
|
| 3. Copy the tokenizer files into the model-containing directory: |
|
|
| ```bash |
| cp checkpoints/$repo_id/tokenizer* out/hf-tinyllama/converted |
| ``` |
|
|
| 4. Run the evaluation harness, for example: |
|
|
| ```bash |
| lm_eval --model hf \ |
| --model_args pretrained=out/hf-tinyllama/converted \ |
| --tasks "hellaswag,gsm8k,truthfulqa_mc2,mmlu,winogrande,arc_challenge" \ |
| --device "cuda:0" \ |
| --batch_size 4 |
| ``` |
|
|