gpt4all/gpt4all-chat/metadata/release.json

556 lines
16 KiB
JSON
Raw Normal View History

[
{
"version": "2.2.2",
"notes":
"
* repeat penalty for both gptj and llama models
* scroll the context window when conversation reaches context limit
* persistent thread count setting
* new default template
* new settings for model path, repeat penalty
* bugfix for settings dialog onEditingFinished
* new tab based settings dialog format
* bugfix for datalake when conversation contains forbidden json chars
* new C library API and split the backend into own separate lib for bindings
* apple signed/notarized dmg installer
* update llama.cpp submodule to latest
* bugfix for too large of a prompt
* support for opt-in only anonymous usage and statistics
* bugfixes for the model downloader and improve performance
* various UI bugfixes and enhancements including the send message textarea automatically wrapping by word
* new startup dialog on first start of a new release displaying release notes and opt-in buttons
* new logo and icons
* fixed apple installer so there is now a symlink in the applications folder
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Matthieu Talbot
* Tim Jobbins
* chad (eachadea)
* Community (beta testers, bug reporters)
"
},
{
"version": "2.3.0",
"notes":
"
* repeat penalty for both gptj and llama models
* scroll the context window when conversation reaches context limit
* persistent thread count setting
* new default template
* new settings for model path, repeat penalty
* bugfix for settings dialog onEditingFinished
* new tab based settings dialog format
* bugfix for datalake when conversation contains forbidden json chars
* new C library API and split the backend into own separate lib for bindings
* apple signed/notarized dmg installer
* update llama.cpp submodule to latest
* bugfix for too large of a prompt
* support for opt-in only anonymous usage and statistics
* bugfixes for the model downloader and improve performance
* various UI bugfixes and enhancements including the send message textarea automatically wrapping by word
* new startup dialog on first start of a new release displaying release notes and opt-in buttons
* new logo and icons
* fixed apple installer so there is now a symlink in the applications folder
* fixed bug with versions
* fixed optout marking
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Matthieu Talbot
* Tim Jobbins
* chad (eachadea)
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.0",
"notes":
"
* reverse prompt for both llama and gptj models which should help stop them from repeating the prompt template
* resumable downloads for models
* chat list in the drawer drop down
* add/remove/rename chats
2023-10-10 18:58:41 +03:00
* persist chats to disk and restore them with full context (WARNING: the average size of each chat on disk is ~1.5GB)
* NOTE: to turn on the persistent chats feature you need to do so via the settings dialog as it is off by default
* automatically rename chats using the AI after the first prompt/response pair
* new usage statistics including more detailed hardware info to help debug problems on older hardware
* fix dialog sizes for those with smaller displays
* add support for persistent contexts and internal model state to the C api
* add a confirm button for deletion of chats
* bugfix for blocking the gui when changing models
* datalake now captures all conversations when network opt-in is turned on
* new much shorter prompt template by default
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.1",
"notes":
"
* compress persistent chats and save order of magnitude disk space on some small chats
* persistent chat files are now stored in same folder as models
* use a thread for deserializing chats on startup so the gui shows window faster
* fail gracefully and early when we detect incompatible hardware
* repeat penalty restore default bugfix
* new mpt backend for mosaic ml's new base model and chat model
* add mpt chat and base model to downloads
* lower memory required for gptj models by using f16 for kv cache
* better error handling for when a model is deleted by user and persistent chat remains
* add a user default model setting so the users preferred model comes up on startup
",
"contributors":
"
* Adam Treat (Nomic AI)
* Zach Nussbaum (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.2",
"notes":
"
* add webserver feature that offers mirror api to chatgpt on localhost:4891
* add chatgpt models installed using openai key to chat client gui
* fixup the memory handling when switching between chats/models to decrease RAM load across the board
* fix bug in thread safety for mpt model and de-duplicated code
* uses compact json format for network
* add remove model option in download dialog
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.3",
"notes":
"
* add webserver feature that offers mirror api to chatgpt on localhost:4891
* add chatgpt models installed using openai key to chat client gui
* fixup the memory handling when switching between chats/models to decrease RAM load across the board
* fix bug in thread safety for mpt model and de-duplicated code
* uses compact json format for network
* add remove model option in download dialog
* remove text-davinci-003 as it is not a chat model
* fix installers on mac and linux to include libllmodel versions
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.4",
"notes":
"
* fix buffer overrun in backend
* bugfix for browse for model directory
* dedup of qml code
* revamp settings dialog UI
* add localdocs plugin (beta) feature allowing scanning of local docs
* various other bugfixes and performance improvements
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller
* Juuso Alasuutari
* Justin Wang
* Community (beta testers, bug reporters)
2023-06-05 19:10:17 +03:00
"
},
{
"version": "2.4.5",
"notes":
"
* bugfix for model download remove
* bugfix for blocking on regenerate
* lots of various ui improvements enhancements
* big new change that brings us up2date with llama.cpp/ggml support for latest models
2023-06-05 21:55:59 +03:00
* advanced avx detection allowing us to fold the two installers into one
* new logging mechanism that allows for bug reports to have more detail
* make localdocs work with server mode
* localdocs fix for stale references after we regenerate
* fix so that browse to dialog on linux
* fix so that you can also just add a path to the textfield
* bugfix for chatgpt and resetting context
* move models.json to github repo so people can pr suggested new models
* allow for new models to be directly downloaded from huggingface in said prs
* better ui for localdocs settings
* better error handling when model fails to load
",
"contributors":
"
* Nils Sauer (Nomic AI)
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Richard Guo (Nomic AI)
* Konstantin Gukov
* Joseph Mearman
* Nandakumar
* Chase McDougall
* mvenditto
* Andriy Mulyar (Nomic AI)
* FoivosC
* Ettore Di Giacinto
* Tim Miller
* Peter Gagarinov
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.6",
"notes":
"
* bugfix for model download remove
* bugfix for blocking on regenerate
* lots of various ui improvements enhancements
* big new change that brings us up2date with llama.cpp/ggml support for latest models
2023-06-05 19:10:17 +03:00
* advanced avx detection allowing us to fold the two installers into one
* new logging mechanism that allows for bug reports to have more detail
* make localdocs work with server mode
* localdocs fix for stale references after we regenerate
* fix so that browse to dialog on linux
* fix so that you can also just add a path to the textfield
* bugfix for chatgpt and resetting context
* move models.json to github repo so people can pr suggested new models
* allow for new models to be directly downloaded from huggingface in said prs
* better ui for localdocs settings
* better error handling when model fails to load
",
"contributors":
"
* Nils Sauer (Nomic AI)
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Richard Guo (Nomic AI)
* Konstantin Gukov
* Joseph Mearman
* Nandakumar
* Chase McDougall
* mvenditto
* Andriy Mulyar (Nomic AI)
* FoivosC
* Ettore Di Giacinto
* Tim Miller
* Peter Gagarinov
* Community (beta testers, bug reporters)
2023-06-16 22:20:23 +03:00
"
},
{
"version": "2.4.7",
"notes":
"
* replit model support
* macos metal accelerated support
* fix markdown for localdocs references
* inline syntax highlighting for python and cpp with more languages coming
* synced with upstream llama.cpp
* ui fixes and default generation settings changes
* backend bugfixes
* allow for loading files directly from huggingface via TheBloke without name changes
",
"contributors":
"
* Nils Sauer (Nomic AI)
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Richard Guo (Nomic AI)
* Andriy Mulyar (Nomic AI)
* Ettore Di Giacinto
* AMOGUS
* Felix Zaslavskiy
* Tim Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.8",
"notes":
"
* replit model support
* macos metal accelerated support
* fix markdown for localdocs references
* inline syntax highlighting for python and cpp with more languages coming
* synced with upstream llama.cpp
* ui fixes and default generation settings changes
* backend bugfixes
* allow for loading files directly from huggingface via TheBloke without name changes
",
"contributors":
"
* Nils Sauer (Nomic AI)
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Richard Guo (Nomic AI)
* Andriy Mulyar (Nomic AI)
* Ettore Di Giacinto
* AMOGUS
* Felix Zaslavskiy
* Tim Miller
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.9",
"notes":
"
* New GPT4All Falcon model
* New Orca models
* Token generation speed is now reported in GUI
* Bugfix for localdocs references when regenerating
* General fixes for thread safety
* Many fixes to UI to add descriptions for error conditions
* Fixes for saving/reloading chats
* Complete refactor of the model download dialog with metadata about models available
* Resume downloads bugfix
* CORS fix
* Documentation fixes and typos
* Latest llama.cpp update
* Update of replit
* Force metal setting
* Fixes for model loading with metal on macOS
",
"contributors":
"
* Nils Sauer (Nomic AI)
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Richard Guo (Nomic AI)
* Andriy Mulyar (Nomic AI)
* cosmic-snow
* AMOGUS
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.10",
"notes":
"
* New GPT4All Falcon model
* New Orca models
* Token generation speed is now reported in GUI
* Bugfix for localdocs references when regenerating
* General fixes for thread safety
* Many fixes to UI to add descriptions for error conditions
* Fixes for saving/reloading chats
* Complete refactor of the model download dialog with metadata about models available
* Resume downloads bugfix
* CORS fix
* Documentation fixes and typos
* Latest llama.cpp update
* Update of replit
* Force metal setting
* Fixes for model loading with metal on macOS
",
"contributors":
"
* Nils Sauer (Nomic AI)
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Richard Guo (Nomic AI)
* Andriy Mulyar (Nomic AI)
* cosmic-snow
* AMOGUS
* Community (beta testers, bug reporters)
2023-07-06 19:25:25 +03:00
"
},
{
"version": "2.4.11",
"notes":
"
* Per model settings
* Character settings
* Adding a system prompt
* Important bugfix for chatgpt install
* Complete refactor and revamp of settings dialog
* New syntax highlighting for java, bash, go
* Use monospace font for syntax highlighting of codeblocks
* New setting for turning off references in localdocs
* Fix memory leaks in falcon model
* Fix for backend memory handling
* Server mode bugfix
* Models.json retrieve bugfix
* Free metal context bugfix
* Add a close dialog feature to all chat dialogs
",
"contributors":
"
* Lakshay Kansal (Nomic AI)
* Matthew Gill
* Brandon Beiler
* cosmic-snow
* Felix Zaslavskiy
* Andriy Mulyar (Nomic AI)
* Aaron Miller (Nomic AI)
* Adam Treat (Nomic AI)
* Community (beta testers, bug reporters)
2023-07-09 20:33:07 +03:00
"
},
{
"version": "2.4.12",
"notes":
"
* Fix bad bug that was breaking numerous current installs (sorry folks!)
* Fix bug with 'browse' button in settings dialog
* Wayland support on linux
* Reduce template ui size in settings dialog
",
"contributors":
"
* Akarshan Biswas
* Adam Treat (Nomic AI)
* Community (beta testers, bug reporters)
"
},
{
"version": "2.4.13",
"notes":
"
* Fix bug with prolonging shutdown with generation
* Fix bug with update model info on deleting chats
* Fix bug with preventing closing of model download dialog
* Allows allow closing the model download dialog
* Fix numerous bugs with download of models.json and provide backup option
* Add json and c# highlighting
* Fix bug with chatgpt crashing
* Fix bug with chatgpt not working for some keys
* Fix bug with mixpanel opt outs not counting
* Fix problem with OOM errors causing crash and then repeating on next start
* Fix default thread setting and provide guardrails
* Fix tap handler in settings dialog for buttons
* Fix color of some text fields on macOS for settings dialog
* Fix problem with startup dialog not closing
* Provide error dialog for settings file not accessible
* Try and fix problems with avx-only detection
* Fix showing error in model downloads unnecessarily
* Prefer 7b models to load by default
* Add Wizard v1.1 to download list
* Rename Orca models to Mini Orca
* Don't use a system prompt unless model was trained with one by default
",
"contributors":
"
* Lakshay Kansal (Nomic AI)
* Aaron Miller (Nomic AI)
* Adam Treat (Nomic AI)
* Community (beta testers, bug reporters)
2023-07-27 18:48:16 +03:00
"
},
{
"version": "2.4.14",
"notes":
"
* Add starcoder model support
* Add ability to switch between light mode/dark mode
* Increase the size of fonts in monospace code blocks a bit
",
"contributors":
"
* Lakshay Kansal (Nomic AI)
* Adam Treat (Nomic AI)
"
},
{
"version": "2.4.15",
"notes":
"
* Add Vulkan GPU backend which allows inference on AMD, Intel and NVIDIA GPUs
* Add ability to switch font sizes
* Various bug fixes
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
* Nils Sauer (Nomic AI)
* Lakshay Kansal (Nomic AI)
"
},
{
"version": "2.4.16",
"notes":
"
* Bugfix for properly falling back to CPU when GPU can't be used
* Report the actual device we're using
* Fix context bugs for GPU accelerated models
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
"
},
{
"version": "2.4.17",
"notes":
"
* Bugfix for properly falling back to CPU when GPU is out of memory
",
"contributors":
"
* Adam Treat (Nomic AI)
* Aaron Miller (Nomic AI)
"
},
{
"version": "2.4.18",
"notes":
"
* Bugfix for devices to show up in the settings combobox on application start and not just on model load
* Send information on requested device and actual device on model load to help assess which model/gpu/os combos are working
",
"contributors":
"
* Adam Treat (Nomic AI)
"
},
{
"version": "2.4.19",
"notes":
"
2023-10-10 18:58:41 +03:00
* Fix a crash on systems with corrupted vulkan drivers or corrupted vulkan dlls
",
"contributors":
"
* Adam Treat (Nomic AI)
"
},
{
"version": "2.5.0",
"notes":
"
* Major new release supports GGUF models only!
* New models like Mistral Instruct, Replit 1.5, Rift Coder and more
* All previous version of ggml-based models are no longer supported
* Extensive changes to vulkan support
* Better GPU error messages
* Prompt processing on the GPU
* Save chats now saves to text (less harddrive space)
* Many more changes
",
"contributors":
"
* Aaron Miller (Nomic AI)
* Jared Van Bortel (Nomic AI)
* Adam Treat (Nomic AI)
* Community (beta testers, bug reporters, bindings authors)
"
}
]