mirror of
https://github.com/tloen/alpaca-lora.git
synced 2024-10-01 01:05:56 -04:00
Add CoLab demo
This commit is contained in:
parent
ec98533876
commit
19af668cb4
@ -1,5 +1,7 @@
|
||||
## 🦙🌲🤏 Alpaca-LoRA: Low-Rank LLaMA Instruct-Tuning
|
||||
|
||||
**Try the pretrained model out on Colab [here](https://colab.research.google.com/drive/1eWAmesrW99p7e1nah5bipn0zikMb8XYC)!**
|
||||
|
||||
This repository contains code for reproducing the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) results using [low-rank adaptation (LoRA)](https://arxiv.org/pdf/2106.09685.pdf).
|
||||
We aim to provide an Instruct model of similar quality to `text-davinci-003` that can run [on a Raspberry Pi](https://twitter.com/miolini/status/1634982361757790209) (for research),
|
||||
but extensions to the `13b`, `30b`, and `65b` models should be feasible with simple changes to the code.
|
||||
@ -12,8 +14,6 @@ as well as Tim Dettmers' [bitsandbytes](https://github.com/TimDettmers/bitsandby
|
||||
|
||||
Without hyperparameter tuning or validation-based checkpointing, the LoRA model produces outputs comparable to the Stanford Alpaca model, though possibly with more minor mistakes. (Please see the outputs included below.) Further tuning might be able to achieve better performance; I invite interested users to give it a try and report their results.
|
||||
|
||||
As usual, I can be reached at https://twitter.com/ecjwg.
|
||||
|
||||
### Setup
|
||||
|
||||
Until Jason Phang's [LLaMA implementation](https://github.com/huggingface/transformers/pull/21955)
|
||||
|
Loading…
Reference in New Issue
Block a user