diff --git a/README.md b/README.md index df4a3f4..32f2c25 100644 --- a/README.md +++ b/README.md @@ -1,8 +1,8 @@ -# alpaca-lora (WIP) +## alpaca-lora (WIP) This repository contains code for reproducing the [Stanford Alpaca results](https://github.com/tatsu-lab/stanford_alpaca#data-release). Users will need to be ready to fork `transformers`. -# Setup +### Setup 1. Install dependencies (**install zphang's transformers fork**) @@ -16,11 +16,11 @@ pip install -q git+https://github.com/huggingface/peft.git 2. [Install bitsandbytes from source](https://github.com/TimDettmers/bitsandbytes/blob/main/compile_from_source.md) -# Inference +### Inference See `generate.py`. This file reads the `decapoda-research/llama-7b-hf` model from the Huggingface model hub and the LoRA weights from `tloen/alpaca-lora-7b`, and runs inference on a specified input. Users should treat this as example code for the use of the model, and modify it as needed. -# Training +### Training Under construction. \ No newline at end of file