mirror of
https://github.com/ravenscroftj/turbopilot.git
synced 2024-10-01 01:06:01 -04:00
185 lines
5.2 KiB
C++
185 lines
5.2 KiB
C++
#include "turbopilot/model.hpp"
|
|
|
|
#include <regex>
|
|
#include <cmath>
|
|
#include <random>
|
|
|
|
|
|
void TurbopilotModel::lock(){
|
|
this->model_lock.lock();
|
|
}
|
|
|
|
void TurbopilotModel::unlock(){
|
|
this->model_lock.unlock();
|
|
}
|
|
|
|
std::stringstream TurbopilotModel::predict(std::string prompt, int max_length, bool include_prompt){
|
|
lock();
|
|
auto result = predict_impl(prompt, max_length, include_prompt);
|
|
unlock();
|
|
return result;
|
|
}
|
|
|
|
void llama_nop(struct ggml_tensor * tensor) { // don't offload by default
|
|
(void) tensor;
|
|
}
|
|
|
|
void gpt_vocab::add_special_token(const std::string & token) {
|
|
special_tokens.push_back(token);
|
|
}
|
|
|
|
|
|
void gpt_split_words(std::string str, std::vector<std::string>& words) {
|
|
const std::string pattern = R"('s|'t|'re|'ve|'m|'ll|'d| ?[[:alpha:]]+| ?[[:digit:]]+| ?[^\s[:alpha:][:digit:]]+|\s+(?!\S)|\s+)";
|
|
const std::regex re(pattern);
|
|
std::smatch m;
|
|
|
|
while (std::regex_search(str, m, re)) {
|
|
for (auto x : m) {
|
|
words.push_back(x);
|
|
}
|
|
str = m.suffix();
|
|
}
|
|
}
|
|
|
|
std::vector<gpt_vocab::id> gpt_tokenize(const gpt_vocab & vocab, const std::string & text) {
|
|
std::vector<std::string> words;
|
|
|
|
// first split the text into words
|
|
{
|
|
std::string str = text;
|
|
|
|
// Generate the subpattern from the special_tokens vector if it's not empty
|
|
if (!vocab.special_tokens.empty()) {
|
|
const std::regex escape(R"([\[\\\^\$\.\|\?\*\+\(\)\{\}])");
|
|
std::string special_tokens_subpattern;
|
|
for (const auto & token : vocab.special_tokens) {
|
|
if (!special_tokens_subpattern.empty()) {
|
|
special_tokens_subpattern += "|";
|
|
}
|
|
special_tokens_subpattern += std::regex_replace(token, escape, R"(\$&)");
|
|
}
|
|
|
|
std::regex re(special_tokens_subpattern);
|
|
std::smatch m;
|
|
// Split the text by special tokens.
|
|
while (std::regex_search(str, m, re)) {
|
|
// Split the substrings in-between special tokens into words.
|
|
gpt_split_words(m.prefix(), words);
|
|
// Add matched special tokens as words.
|
|
for (auto x : m) {
|
|
words.push_back(x);
|
|
}
|
|
str = m.suffix();
|
|
}
|
|
// Remaining text without special tokens will be handled below.
|
|
}
|
|
|
|
gpt_split_words(str, words);
|
|
}
|
|
|
|
// find the longest token that forms each word in words:
|
|
std::vector<gpt_vocab::id> tokens;
|
|
for (const auto & word : words) {
|
|
for (int i = 0; i < (int) word.size(); ){
|
|
for (int j = word.size() - 1; j >= i; j--){
|
|
auto cand = word.substr(i, j-i+1);
|
|
auto it = vocab.token_to_id.find(cand);
|
|
if (it != vocab.token_to_id.end()){ // word.substr(i, j-i+1) in vocab
|
|
tokens.push_back(it->second);
|
|
i = j + 1;
|
|
break;
|
|
}
|
|
else if (j == i){ // word.substr(i, 1) has no matching
|
|
fprintf(stderr, "%s: unknown token '%s'\n", __func__, word.substr(i, 1).data());
|
|
i++;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return tokens;
|
|
}
|
|
|
|
gpt_vocab::id gpt_sample_top_k_top_p(
|
|
const gpt_vocab & vocab,
|
|
const float * logits,
|
|
int top_k,
|
|
double top_p,
|
|
double temp,
|
|
std::mt19937 & rng) {
|
|
int n_logits = vocab.id_to_token.size();
|
|
|
|
std::vector<std::pair<double, gpt_vocab::id>> logits_id;
|
|
logits_id.reserve(n_logits);
|
|
|
|
{
|
|
const double scale = 1.0/temp;
|
|
for (int i = 0; i < n_logits; ++i) {
|
|
logits_id.push_back(std::make_pair(logits[i]*scale, i));
|
|
}
|
|
}
|
|
|
|
// find the top K tokens
|
|
std::partial_sort(
|
|
logits_id.begin(),
|
|
logits_id.begin() + top_k, logits_id.end(),
|
|
[](const std::pair<double, gpt_vocab::id> & a, const std::pair<double, gpt_vocab::id> & b) {
|
|
return a.first > b.first;
|
|
});
|
|
|
|
logits_id.resize(top_k);
|
|
|
|
double maxl = -INFINITY;
|
|
for (const auto & kv : logits_id) {
|
|
maxl = std::max(maxl, kv.first);
|
|
}
|
|
|
|
// compute probs for the top K tokens
|
|
std::vector<double> probs;
|
|
probs.reserve(logits_id.size());
|
|
|
|
double sum = 0.0;
|
|
for (const auto & kv : logits_id) {
|
|
double p = exp(kv.first - maxl);
|
|
probs.push_back(p);
|
|
sum += p;
|
|
}
|
|
|
|
// normalize the probs
|
|
for (auto & p : probs) {
|
|
p /= sum;
|
|
}
|
|
|
|
if (top_p < 1.0f) {
|
|
double cumsum = 0.0f;
|
|
for (int i = 0; i < top_k; i++) {
|
|
cumsum += probs[i];
|
|
if (cumsum >= top_p) {
|
|
top_k = i + 1;
|
|
probs.resize(top_k);
|
|
logits_id.resize(top_k);
|
|
break;
|
|
}
|
|
}
|
|
|
|
cumsum = 1.0/cumsum;
|
|
for (int i = 0; i < (int) probs.size(); i++) {
|
|
probs[i] *= cumsum;
|
|
}
|
|
}
|
|
|
|
//printf("\n");
|
|
//for (int i = 0; i < (int) probs.size(); i++) {
|
|
// printf("%d: '%s' %f\n", i, vocab.id_to_token.at(logits_id[i].second).c_str(), probs[i]);
|
|
//}
|
|
//exit(0);
|
|
|
|
std::discrete_distribution<> dist(probs.begin(), probs.end());
|
|
int idx = dist(rng);
|
|
|
|
return logits_id[idx].second;
|
|
}
|
|
|
|
|