mirror of
https://github.com/ollama/ollama.git
synced 2026-04-27 11:15:40 +02:00
Compare commits
4 Commits
jmorganca/
...
parth/remo
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
7b152860c2 | ||
|
|
467bbc0dd5 | ||
|
|
6d9f9323c5 | ||
|
|
0c2489605d |
2
.gitattributes
vendored
2
.gitattributes
vendored
@@ -19,6 +19,8 @@ ml/backend/**/*.comp linguist-vendored
|
|||||||
ml/backend/**/*.glsl linguist-vendored
|
ml/backend/**/*.glsl linguist-vendored
|
||||||
ml/backend/**/CMakeLists.txt linguist-vendored
|
ml/backend/**/CMakeLists.txt linguist-vendored
|
||||||
|
|
||||||
|
app/webview linguist-vendored
|
||||||
|
|
||||||
llama/build-info.cpp linguist-generated
|
llama/build-info.cpp linguist-generated
|
||||||
ml/backend/ggml/ggml/src/ggml-metal/ggml-metal-embed.s linguist-generated
|
ml/backend/ggml/ggml/src/ggml-metal/ggml-metal-embed.s linguist-generated
|
||||||
|
|
||||||
|
|||||||
@@ -23,7 +23,6 @@ linters:
|
|||||||
- nolintlint
|
- nolintlint
|
||||||
- nosprintfhostport
|
- nosprintfhostport
|
||||||
- perfsprint
|
- perfsprint
|
||||||
- prealloc
|
|
||||||
- sloglint
|
- sloglint
|
||||||
- staticcheck
|
- staticcheck
|
||||||
- unconvert
|
- unconvert
|
||||||
|
|||||||
@@ -98,6 +98,9 @@ func GPUDevices(ctx context.Context, runners []ml.FilteredRunnerDiscovery) []ml.
|
|||||||
continue
|
continue
|
||||||
} else if jetpack != "" && filepath.Base(dir) != "cuda_"+jetpack {
|
} else if jetpack != "" && filepath.Base(dir) != "cuda_"+jetpack {
|
||||||
continue
|
continue
|
||||||
|
} else if jetpack == "" && strings.Contains(filepath.Base(dir), "cuda_jetpack") {
|
||||||
|
slog.Debug("jetpack not detected (set JETSON_JETPACK or OLLAMA_LLM_LIBRARY to override), skipping", "libDir", dir)
|
||||||
|
continue
|
||||||
} else if !envconfig.EnableVulkan() && strings.Contains(filepath.Base(dir), "vulkan") {
|
} else if !envconfig.EnableVulkan() && strings.Contains(filepath.Base(dir), "vulkan") {
|
||||||
slog.Info("experimental Vulkan support disabled. To enable, set OLLAMA_VULKAN=1")
|
slog.Info("experimental Vulkan support disabled. To enable, set OLLAMA_VULKAN=1")
|
||||||
continue
|
continue
|
||||||
|
|||||||
11
docs/faq.mdx
11
docs/faq.mdx
@@ -57,8 +57,13 @@ ollama ps
|
|||||||
```
|
```
|
||||||
|
|
||||||
<Info>
|
<Info>
|
||||||
**Output**: ``` NAME ID SIZE PROCESSOR UNTIL llama3:70b bcfb190ca3a7 42 GB
|
|
||||||
100% GPU 4 minutes from now ```
|
**Output**:
|
||||||
|
|
||||||
|
```
|
||||||
|
NAME ID SIZE PROCESSOR UNTIL
|
||||||
|
llama3:70b bcfb190ca3a7 42 GB 100% GPU 4 minutes from now
|
||||||
|
```
|
||||||
</Info>
|
</Info>
|
||||||
|
|
||||||
The `Processor` column will show which memory the model was loaded in to:
|
The `Processor` column will show which memory the model was loaded in to:
|
||||||
@@ -385,4 +390,4 @@ Ollama for Windows and macOS register as a login item during installation. You
|
|||||||
- In `Task Manager` go to the `Startup apps` tab, search for `ollama` then click `Disable`
|
- In `Task Manager` go to the `Startup apps` tab, search for `ollama` then click `Disable`
|
||||||
|
|
||||||
**MacOS**
|
**MacOS**
|
||||||
- Open `Settings` and search for "Login Items", find the `Ollama` entry under "Allow in the Background`, then click the slider to disable.
|
- Open `Settings` and search for "Login Items", find the `Ollama` entry under "Allow in the Background`, then click the slider to disable.
|
||||||
|
|||||||
Reference in New Issue
Block a user