From ea0e96b0170d3533de0766505a12bc99dfc6c26f Mon Sep 17 00:00:00 2001 From: Chris Alexiuk <114439245+chris-alexiuk@users.noreply.github.com> Date: Thu, 13 Apr 2023 16:46:48 -0400 Subject: [PATCH] Added reference to Dolly 15k Dataset (#330) * Update README.md * Feature HF dataset more prominently --------- Co-authored-by: Eric Wang --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index fd5347c..7e37de7 100644 --- a/README.md +++ b/README.md @@ -155,6 +155,7 @@ docker-compose down --volumes --rmi all - [Alpaca-LoRA-Serve](https://github.com/deep-diver/Alpaca-LoRA-Serve), a ChatGPT-style interface for Alpaca models - [AlpacaDataCleaned](https://github.com/gururise/AlpacaDataCleaned), a project to improve the quality of the Alpaca dataset - [GPT-4 Alpaca Data](https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM) a project to port synthetic data creation to GPT-4 +- [dolly-15k-instruction-alpaca-format](https://huggingface.co/datasets/c-s-ale/dolly-15k-instruction-alpaca-format), an Alpaca-compatible version of [Databricks' Dolly 15k human-generated instruct dataset](https://github.com/databrickslabs/dolly/tree/master/data) (see [blog](https://www.databricks.com/blog/2023/04/12/dolly-first-open-commercially-viable-instruction-tuned-llm)) - Various adapter weights (download at own risk): - 7B: -