mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-11-05 04:12:29 +03:00
48 lines
1.6 KiB
Markdown
48 lines
1.6 KiB
Markdown
|
## QA Checklist
|
||
|
|
||
|
1. Ensure you have a fresh install by backing up and then deleting the following:
|
||
|
|
||
|
### Windows
|
||
|
* Settings directory: TBD
|
||
|
* Models directory: TBD
|
||
|
### Mac
|
||
|
* Settings directory: TBD
|
||
|
* Models directory: TBD
|
||
|
### Linux
|
||
|
* Settings directory: TBD
|
||
|
* Models directory: TBD
|
||
|
|
||
|
^ Note: If you've changed your models directory manually via the settings you need to backup and delete that one
|
||
|
|
||
|
2. Go through every view and ensure that things display correctly and familiarize yourself with the application flow
|
||
|
|
||
|
3. Navigate to the models view and download Llama 3 Instruct
|
||
|
|
||
|
4. Navigate to the models view and search for "TheBloke mistral 7b" and download "TheBloke/Mistral-7B-Instruct-v0.1-GGUF"
|
||
|
|
||
|
5. Navigate to the chat view and open new chats and load these models
|
||
|
|
||
|
6. Chat with the models and exercise them. Rename the chats. Delete chats. Open new chats. Switch models when in chats.
|
||
|
|
||
|
7. Create a new localdocs collection from a directory of .txt or .pdf files on your hard drive
|
||
|
|
||
|
8. Enable the new collection in chats (especially with Llama 3 Instruct) and exercise the localdocs feature
|
||
|
|
||
|
9. Go to the settings view and explore each setting
|
||
|
|
||
|
10. Remove collections in localdocs and re-add them. Rebuild collections.
|
||
|
|
||
|
11. Try to break the app
|
||
|
|
||
|
EXTRA CREDIT:
|
||
|
|
||
|
1. If you have a openai api key install GPT-4 model and chat with it
|
||
|
|
||
|
2. If you have a nomic api key install the remote nomic embedding model for localdocs (see if you can discover how to do this)
|
||
|
|
||
|
3. If you have a python script that targets openai API then enable server mode and try this
|
||
|
|
||
|
4. Really try and break the app
|
||
|
|
||
|
All feedback is welcome
|