mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
make scripts executable (#1555)
This commit is contained in:
parent
f414c28589
commit
e90263c23f
@ -1 +1 @@
|
||||
Subproject commit a8ed8c858985ef94d97a3cf2c97085b680c6d5d0
|
||||
Subproject commit 2dee60214b0001cf03e1cec0a53a61a17b55c1eb
|
1
gpt4all-bindings/cli/app.py
Normal file → Executable file
1
gpt4all-bindings/cli/app.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
"""GPT4All CLI
|
||||
|
||||
The GPT4All CLI is a self-contained script based on the `gpt4all` and `typer` packages. It offers a
|
||||
|
@ -1,3 +1,4 @@
|
||||
#!/bin/sh
|
||||
mkdir -p runtimes
|
||||
rm -rf runtimes/linux-x64
|
||||
mkdir -p runtimes/linux-x64/native
|
||||
|
1
gpt4all-bindings/python/gpt4all/tests/test_embed_timings.py
Normal file → Executable file
1
gpt4all-bindings/python/gpt4all/tests/test_embed_timings.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import sys
|
||||
import time
|
||||
from io import StringIO
|
||||
|
0
gpt4all-bindings/typescript/scripts/build_unix.sh
Normal file → Executable file
0
gpt4all-bindings/typescript/scripts/build_unix.sh
Normal file → Executable file
1
gpt4all-chat/cmake/sign_dmg.py
Normal file → Executable file
1
gpt4all-chat/cmake/sign_dmg.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import os
|
||||
import subprocess
|
||||
import tempfile
|
||||
|
3
gpt4all-training/build_map.py
Normal file → Executable file
3
gpt4all-training/build_map.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import numpy as np
|
||||
from nomic import atlas
|
||||
import glob
|
||||
@ -51,4 +52,4 @@ atlas.map_embeddings(embeddings,
|
||||
colorable_fields=["source", "loss", "trained_on"],
|
||||
build_topic_model=True,
|
||||
topic_label_field="inputs",
|
||||
reset_project_if_exists=True,)
|
||||
reset_project_if_exists=True,)
|
||||
|
3
gpt4all-training/clean.py
Normal file → Executable file
3
gpt4all-training/clean.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import numpy as np
|
||||
import glob
|
||||
import os
|
||||
@ -71,4 +72,4 @@ for file in glob.glob(os.path.join(prompt_generation_dir, "*.jsonl")):
|
||||
|
||||
clean_name = file.split(".jsonl")[0] + "_clean.jsonl"
|
||||
print(f"writing to {curr_len} rows to {clean_name}")
|
||||
df.to_json(clean_name, orient="records", lines=True)
|
||||
df.to_json(clean_name, orient="records", lines=True)
|
||||
|
0
gpt4all-training/create_hostname.sh
Normal file → Executable file
0
gpt4all-training/create_hostname.sh
Normal file → Executable file
1
gpt4all-training/eval_figures.py
Normal file → Executable file
1
gpt4all-training/eval_figures.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import glob
|
||||
import pickle
|
||||
import numpy as np
|
||||
|
1
gpt4all-training/eval_self_instruct.py
Normal file → Executable file
1
gpt4all-training/eval_self_instruct.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import json
|
||||
import torch
|
||||
import pickle
|
||||
|
1
gpt4all-training/generate.py
Normal file → Executable file
1
gpt4all-training/generate.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||
from peft import PeftModelForCausalLM
|
||||
from read import read_config
|
||||
|
1
gpt4all-training/inference.py
Normal file → Executable file
1
gpt4all-training/inference.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||
import torch
|
||||
import torch.nn as nn
|
||||
|
0
gpt4all-training/launcher.sh
Normal file → Executable file
0
gpt4all-training/launcher.sh
Normal file → Executable file
1
gpt4all-training/train.py
Normal file → Executable file
1
gpt4all-training/train.py
Normal file → Executable file
@ -1,3 +1,4 @@
|
||||
#!/usr/bin/env python3
|
||||
import os
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer, get_scheduler
|
||||
import torch
|
||||
|
Loading…
Reference in New Issue
Block a user