mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-11-05 04:12:29 +03:00
9273b49b62
Signed-off-by: Adam Treat <treat.adam@gmail.com> Signed-off-by: Jared Van Bortel <jared@nomic.ai> Co-authored-by: Jared Van Bortel <jared@nomic.ai>
81 lines
2.9 KiB
C++
81 lines
2.9 KiB
C++
#include "chatlistmodel.h"
|
|
#include "config.h"
|
|
#include "download.h"
|
|
#include "llm.h"
|
|
#include "localdocs.h"
|
|
#include "logger.h"
|
|
#include "modellist.h"
|
|
#include "mysettings.h"
|
|
#include "network.h"
|
|
|
|
#include "../gpt4all-backend/llmodel.h"
|
|
|
|
#include <QCoreApplication>
|
|
#include <QGuiApplication>
|
|
#include <QObject>
|
|
#include <QQmlApplicationEngine>
|
|
#include <QQmlEngine>
|
|
#include <QSettings>
|
|
#include <QString>
|
|
#include <QUrl>
|
|
#include <Qt>
|
|
|
|
int main(int argc, char *argv[])
|
|
{
|
|
QCoreApplication::setOrganizationName("nomic.ai");
|
|
QCoreApplication::setOrganizationDomain("gpt4all.io");
|
|
QCoreApplication::setApplicationName("GPT4All");
|
|
QCoreApplication::setApplicationVersion(APP_VERSION);
|
|
QSettings::setDefaultFormat(QSettings::IniFormat);
|
|
|
|
Logger::globalInstance();
|
|
|
|
QGuiApplication app(argc, argv);
|
|
QQmlApplicationEngine engine;
|
|
|
|
QString llmodelSearchPaths = QCoreApplication::applicationDirPath();
|
|
const QString libDir = QCoreApplication::applicationDirPath() + "/../lib/";
|
|
if (LLM::directoryExists(libDir))
|
|
llmodelSearchPaths += ";" + libDir;
|
|
#if defined(Q_OS_MAC)
|
|
const QString binDir = QCoreApplication::applicationDirPath() + "/../../../";
|
|
if (LLM::directoryExists(binDir))
|
|
llmodelSearchPaths += ";" + binDir;
|
|
const QString frameworksDir = QCoreApplication::applicationDirPath() + "/../Frameworks/";
|
|
if (LLM::directoryExists(frameworksDir))
|
|
llmodelSearchPaths += ";" + frameworksDir;
|
|
#endif
|
|
LLModel::Implementation::setImplementationsSearchPath(llmodelSearchPaths.toStdString());
|
|
|
|
qmlRegisterSingletonInstance("mysettings", 1, 0, "MySettings", MySettings::globalInstance());
|
|
qmlRegisterSingletonInstance("modellist", 1, 0, "ModelList", ModelList::globalInstance());
|
|
qmlRegisterSingletonInstance("chatlistmodel", 1, 0, "ChatListModel", ChatListModel::globalInstance());
|
|
qmlRegisterSingletonInstance("llm", 1, 0, "LLM", LLM::globalInstance());
|
|
qmlRegisterSingletonInstance("download", 1, 0, "Download", Download::globalInstance());
|
|
qmlRegisterSingletonInstance("network", 1, 0, "Network", Network::globalInstance());
|
|
qmlRegisterSingletonInstance("localdocs", 1, 0, "LocalDocs", LocalDocs::globalInstance());
|
|
const QUrl url(u"qrc:/gpt4all/main.qml"_qs);
|
|
|
|
QObject::connect(&engine, &QQmlApplicationEngine::objectCreated,
|
|
&app, [url](QObject *obj, const QUrl &objUrl) {
|
|
if (!obj && url == objUrl)
|
|
QCoreApplication::exit(-1);
|
|
}, Qt::QueuedConnection);
|
|
engine.load(url);
|
|
|
|
#if 0
|
|
QDirIterator it("qrc:", QDirIterator::Subdirectories);
|
|
while (it.hasNext()) {
|
|
qDebug() << it.next();
|
|
}
|
|
#endif
|
|
|
|
int res = app.exec();
|
|
|
|
// Make sure ChatLLM threads are joined before global destructors run.
|
|
// Otherwise, we can get a heap-use-after-free inside of llama.cpp.
|
|
ChatListModel::globalInstance()->destroyChats();
|
|
|
|
return res;
|
|
}
|