mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
45 lines
1.0 KiB
C++
45 lines
1.0 KiB
C++
#ifndef LLM_H
|
|
#define LLM_H
|
|
|
|
#include <QObject>
|
|
|
|
#include "chatlistmodel.h"
|
|
|
|
class LLM : public QObject
|
|
{
|
|
Q_OBJECT
|
|
Q_PROPERTY(ChatListModel *chatListModel READ chatListModel NOTIFY chatListModelChanged)
|
|
Q_PROPERTY(int32_t threadCount READ threadCount WRITE setThreadCount NOTIFY threadCountChanged)
|
|
Q_PROPERTY(bool compatHardware READ compatHardware NOTIFY compatHardwareChanged)
|
|
|
|
public:
|
|
static LLM *globalInstance();
|
|
|
|
ChatListModel *chatListModel() const { return m_chatListModel; }
|
|
int32_t threadCount() const;
|
|
void setThreadCount(int32_t n_threads);
|
|
bool compatHardware() const { return m_compatHardware; }
|
|
|
|
Q_INVOKABLE bool checkForUpdates() const;
|
|
|
|
Q_SIGNALS:
|
|
void chatListModelChanged();
|
|
void threadCountChanged();
|
|
void compatHardwareChanged();
|
|
|
|
private Q_SLOTS:
|
|
void aboutToQuit();
|
|
|
|
private:
|
|
ChatListModel *m_chatListModel;
|
|
int32_t m_threadCount;
|
|
bool m_compatHardware;
|
|
|
|
private:
|
|
explicit LLM();
|
|
~LLM() {}
|
|
friend class MyLLM;
|
|
};
|
|
|
|
#endif // LLM_H
|