gpt4all/gpt4all-chat/llm.h

50 lines
1.3 KiB
C
Raw Normal View History

2023-04-09 03:28:39 +00:00
#ifndef LLM_H
#define LLM_H
#include <QObject>
2023-05-01 21:13:20 +00:00
#include "chatlistmodel.h"
2023-04-09 03:28:39 +00:00
class LLM : public QObject
{
Q_OBJECT
2023-05-01 21:13:20 +00:00
Q_PROPERTY(ChatListModel *chatListModel READ chatListModel NOTIFY chatListModelChanged)
Q_PROPERTY(int32_t threadCount READ threadCount WRITE setThreadCount NOTIFY threadCountChanged)
2023-05-11 20:46:25 +00:00
Q_PROPERTY(bool serverEnabled READ serverEnabled WRITE setServerEnabled NOTIFY serverEnabledChanged)
Q_PROPERTY(bool compatHardware READ compatHardware NOTIFY compatHardwareChanged)
2023-04-09 03:28:39 +00:00
public:
static LLM *globalInstance();
2023-05-01 21:13:20 +00:00
ChatListModel *chatListModel() const { return m_chatListModel; }
int32_t threadCount() const;
void setThreadCount(int32_t n_threads);
2023-05-11 20:46:25 +00:00
bool serverEnabled() const;
void setServerEnabled(bool enabled);
bool compatHardware() const { return m_compatHardware; }
Q_INVOKABLE bool checkForUpdates() const;
Q_INVOKABLE bool directoryExists(const QString &path) const;
Q_INVOKABLE bool fileExists(const QString &path) const;
2023-04-09 03:28:39 +00:00
Q_SIGNALS:
2023-05-01 21:13:20 +00:00
void chatListModelChanged();
void threadCountChanged();
2023-05-11 20:46:25 +00:00
void serverEnabledChanged();
void compatHardwareChanged();
2023-04-09 03:28:39 +00:00
private:
2023-05-01 21:13:20 +00:00
ChatListModel *m_chatListModel;
int32_t m_threadCount;
2023-05-11 20:46:25 +00:00
bool m_serverEnabled;
bool m_compatHardware;
2023-04-09 03:28:39 +00:00
private:
explicit LLM();
~LLM() {}
friend class MyLLM;
};
#endif // LLM_H