mirror of
https://github.com/tloen/alpaca-lora.git
synced 2024-10-01 01:05:56 -04:00
To-dos, etc
This commit is contained in:
parent
6707775517
commit
a7ccc3603e
13
README.md
13
README.md
@ -1,4 +1,4 @@
|
|||||||
## alpaca-lora (WIP)
|
## 🦙🌲🤏 Alpaca (Low-Rank Edition)
|
||||||
|
|
||||||
This repository contains code for reproducing the [Stanford Alpaca results](https://github.com/tatsu-lab/stanford_alpaca#data-release). Users will need to be ready to fork `transformers`.
|
This repository contains code for reproducing the [Stanford Alpaca results](https://github.com/tatsu-lab/stanford_alpaca#data-release). Users will need to be ready to fork `transformers`.
|
||||||
|
|
||||||
@ -23,4 +23,13 @@ See `generate.py`. This file reads the `decapoda-research/llama-7b-hf` model fro
|
|||||||
|
|
||||||
### Training
|
### Training
|
||||||
|
|
||||||
Under construction.
|
Under construction.
|
||||||
|
|
||||||
|
|
||||||
|
### To do
|
||||||
|
|
||||||
|
- [ ] Hyperparameter tuning
|
||||||
|
- [ ] Documentation for notebook
|
||||||
|
- [ ] Support for `13b`, `30b`, `65b`
|
||||||
|
- [ ] Inference CLI and evaluation
|
||||||
|
- [ ] Better disclaimers about why using LLaMA without permission is very bad!
|
Loading…
Reference in New Issue
Block a user