mirror of
https://github.com/juspay/services-flake.git
synced 2024-09-17 15:28:33 +03:00
chore(docs): ollama: better intro, and GPU pointers
This commit is contained in:
parent
1ab4a8a9aa
commit
4cb3824de9
@ -1,6 +1,6 @@
|
|||||||
# Ollama
|
# Ollama
|
||||||
|
|
||||||
[Ollama](https://github.com/ollama/ollama) enables you to get up and running with Llama 3, Mistral, Gemma, and other large language models.
|
[Ollama](https://github.com/ollama/ollama) enables you to easily run large language models (LLMs) locally. It supports Llama 3, Mistral, Gemma and [many others](https://ollama.com/library).
|
||||||
|
|
||||||
## Getting Started
|
## Getting Started
|
||||||
|
|
||||||
@ -15,7 +15,9 @@
|
|||||||
|
|
||||||
By default Ollama uses the CPU for inference. To enable GPU acceleration:
|
By default Ollama uses the CPU for inference. To enable GPU acceleration:
|
||||||
|
|
||||||
### Cuda
|
### CUDA
|
||||||
|
|
||||||
|
For NVIDIA GPUs.
|
||||||
|
|
||||||
```nix
|
```nix
|
||||||
# In `perSystem.process-compose.<name>`
|
# In `perSystem.process-compose.<name>`
|
||||||
@ -29,6 +31,8 @@ By default Ollama uses the CPU for inference. To enable GPU acceleration:
|
|||||||
|
|
||||||
### ROCm
|
### ROCm
|
||||||
|
|
||||||
|
For Radeon GPUs.
|
||||||
|
|
||||||
```nix
|
```nix
|
||||||
# In `perSystem.process-compose.<name>`
|
# In `perSystem.process-compose.<name>`
|
||||||
{
|
{
|
||||||
|
Loading…
Reference in New Issue
Block a user