url
stringlengths 51
54
| repository_url
stringclasses 1
value | labels_url
stringlengths 65
68
| comments_url
stringlengths 60
63
| events_url
stringlengths 58
61
| html_url
stringlengths 39
44
| id
int64 1.78B
2.82B
| node_id
stringlengths 18
19
| number
int64 1
8.69k
| title
stringlengths 1
382
| user
dict | labels
listlengths 0
5
| state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
listlengths 0
2
| milestone
null | comments
int64 0
323
| created_at
timestamp[s] | updated_at
timestamp[s] | closed_at
timestamp[s] | author_association
stringclasses 4
values | sub_issues_summary
dict | active_lock_reason
null | draft
bool 2
classes | pull_request
dict | body
stringlengths 2
118k
⌀ | closed_by
dict | reactions
dict | timeline_url
stringlengths 60
63
| performed_via_github_app
null | state_reason
stringclasses 4
values | is_pull_request
bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/5614
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5614/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5614/comments
|
https://api.github.com/repos/ollama/ollama/issues/5614/events
|
https://github.com/ollama/ollama/pull/5614
| 2,401,766,631
|
PR_kwDOJ0Z1Ps51BVsT
| 5,614
|
OpenAI: Support Tools
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-07-10T21:21:37
| 2024-07-23T23:02:27
| 2024-07-17T03:52:59
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5614",
"html_url": "https://github.com/ollama/ollama/pull/5614",
"diff_url": "https://github.com/ollama/ollama/pull/5614.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5614.patch",
"merged_at": "2024-07-17T03:52:59"
}
|
```
curl -s 127.0.0.1:11434/v1/chat/completions -d '{
"model": "mike/mistral",
"messages": [
{
"role": "user",
"content": "What'\''s the weather like today in Paris?"
},
{
"role": "assistant",
"tool_calls": [
{
"id": "89a1e453-0bce-4de3-a456-c54bed09c520",
"type": "function",
"function": {
"name": "get_current_weather",
"arguments": "{\"location\": \"Paris, France\", \"format\": \"celsius\"}"
}
}
]
},
{
"role": "tool",
"tool_call_id": "89a1e453-0bce-4de3-a456-c54bed09c520",
"content": "22"
},
{
"role": "assistant",
"content": "The weather in Paris is 22 degrees celsius."
},
{
"role": "user",
"content": "What'\''s the weather like today in San Francisco and Toronto?"
}
],
"tools": [
{
"type": "function",
"function": {
"name": "get_current_weather",
"description": "Get the current weather",
"parameters": {
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "The city and state, e.g. San Francisco, CA"
},
"format": {
"type": "string",
"enum": [
"celsius",
"fahrenheit"
],
"description": "The temperature unit to use. Infer this from the users location."
}
},
"required": [
"location",
"format"
]
}
}
}
],
"stream": false,
"temperature": 0
}' | jq
```
```
{
"id": "chatcmpl-920",
"object": "chat.completion",
"created": 1721187203,
"model": "mike/mistral",
"system_fingerprint": "fp_ollama",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "",
"tool_calls": [
{
"id": "call_q9jtgbw2",
"type": "function",
"function": {
"name": "get_current_weather",
"arguments": "{\"format\":\"celsius\",\"location\":\"San Francisco, CA\"}"
}
},
{
"id": "call_9chg85cd",
"type": "function",
"function": {
"name": "get_current_weather",
"arguments": "{\"format\":\"celsius\",\"location\":\"Toronto, Canada\"}"
}
}
]
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 198,
"completion_tokens": 65,
"total_tokens": 263
}
}
```
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5614/reactions",
"total_count": 8,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 8,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5614/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/700
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/700/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/700/comments
|
https://api.github.com/repos/ollama/ollama/issues/700/events
|
https://github.com/ollama/ollama/pull/700
| 1,926,883,924
|
PR_kwDOJ0Z1Ps5b77h7
| 700
|
rename server subprocess
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-10-04T19:53:55
| 2023-10-06T14:15:43
| 2023-10-06T14:15:42
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/700",
"html_url": "https://github.com/ollama/ollama/pull/700",
"diff_url": "https://github.com/ollama/ollama/pull/700.diff",
"patch_url": "https://github.com/ollama/ollama/pull/700.patch",
"merged_at": "2023-10-06T14:15:42"
}
|
rename llama.cpp `server.exe` to `ollama-runner`. This makes it easier to see that the subprocess is associated with ollama.
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/700/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/700/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8647
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8647/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8647/comments
|
https://api.github.com/repos/ollama/ollama/issues/8647/events
|
https://github.com/ollama/ollama/issues/8647
| 2,817,099,829
|
I_kwDOJ0Z1Ps6n6Yg1
| 8,647
|
ollama[2075]: cuda driver library failed to get device context 999time=2025-01-29T03:21:43.859+01:00 level=WARN source=gpu.go:449 msg="error looking up nvidia GPU memory"
|
{
"login": "kurdo2000",
"id": 137941417,
"node_id": "U_kgDOCDjRqQ",
"avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kurdo2000",
"html_url": "https://github.com/kurdo2000",
"followers_url": "https://api.github.com/users/kurdo2000/followers",
"following_url": "https://api.github.com/users/kurdo2000/following{/other_user}",
"gists_url": "https://api.github.com/users/kurdo2000/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kurdo2000/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kurdo2000/subscriptions",
"organizations_url": "https://api.github.com/users/kurdo2000/orgs",
"repos_url": "https://api.github.com/users/kurdo2000/repos",
"events_url": "https://api.github.com/users/kurdo2000/events{/privacy}",
"received_events_url": "https://api.github.com/users/kurdo2000/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
}
] |
closed
| false
| null |
[] | null | 2
| 2025-01-29T02:34:23
| 2025-01-30T03:09:02
| 2025-01-30T03:09:01
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hi,
is my GPU not supported or what's the error?
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.120 Driver Version: 550.120 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce GTX 1660 Off | 00000000:01:00.0 On | N/A |
| 0% 48C P0 17W / 130W | 1602MiB / 6144MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| 0 N/A N/A 2627 G /usr/lib/xorg/Xorg 1032MiB |
| 0 N/A N/A 2853 G /usr/bin/gnome-shell 79MiB |
| 0 N/A N/A 3461 G /usr/libexec/xdg-desktop-portal-gnome 108MiB |
| 0 N/A N/A 4054 G ...irefox/4793/usr/lib/firefox/firefox 203MiB |
| 0 N/A N/A 25159 G /usr/bin/nautilus 40MiB |
| 0 N/A N/A 106416 G ...erProcess --variations-seed-version 82MiB |
+-----------------------------------------------------------------------------------------+
nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2024 NVIDIA Corporation
Built on Tue_Oct_29_23:50:19_PDT_2024
Cuda compilation tools, release 12.6, V12.6.85
Build cuda_12.6.r12.6/compiler.35059454_0
If you still need more context, please ask. I'm new here.
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.5.7
|
{
"login": "kurdo2000",
"id": 137941417,
"node_id": "U_kgDOCDjRqQ",
"avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kurdo2000",
"html_url": "https://github.com/kurdo2000",
"followers_url": "https://api.github.com/users/kurdo2000/followers",
"following_url": "https://api.github.com/users/kurdo2000/following{/other_user}",
"gists_url": "https://api.github.com/users/kurdo2000/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kurdo2000/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kurdo2000/subscriptions",
"organizations_url": "https://api.github.com/users/kurdo2000/orgs",
"repos_url": "https://api.github.com/users/kurdo2000/repos",
"events_url": "https://api.github.com/users/kurdo2000/events{/privacy}",
"received_events_url": "https://api.github.com/users/kurdo2000/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8647/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8647/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8067
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8067/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8067/comments
|
https://api.github.com/repos/ollama/ollama/issues/8067/events
|
https://github.com/ollama/ollama/issues/8067
| 2,735,177,482
|
I_kwDOJ0Z1Ps6jB38K
| 8,067
|
have a ques
|
{
"login": "17692019663",
"id": 57002321,
"node_id": "MDQ6VXNlcjU3MDAyMzIx",
"avatar_url": "https://avatars.githubusercontent.com/u/57002321?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/17692019663",
"html_url": "https://github.com/17692019663",
"followers_url": "https://api.github.com/users/17692019663/followers",
"following_url": "https://api.github.com/users/17692019663/following{/other_user}",
"gists_url": "https://api.github.com/users/17692019663/gists{/gist_id}",
"starred_url": "https://api.github.com/users/17692019663/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/17692019663/subscriptions",
"organizations_url": "https://api.github.com/users/17692019663/orgs",
"repos_url": "https://api.github.com/users/17692019663/repos",
"events_url": "https://api.github.com/users/17692019663/events{/privacy}",
"received_events_url": "https://api.github.com/users/17692019663/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
},
{
"id": 6677677816,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgVG-A",
"url": "https://api.github.com/repos/ollama/ollama/labels/docker",
"name": "docker",
"color": "0052CC",
"default": false,
"description": "Issues relating to using ollama in containers"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-12-12T08:17:34
| 2024-12-23T08:12:10
| 2024-12-23T08:12:10
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
C:\Users\Administrator> docker exec -it ollama ollama run llama3.2:1b
pulling manifest
pulling 74701a8c35f6... 100% ▕████████████████████████████████████████████████████████▏ 1.3 GB
pulling 966de95ca8a6... 100% ▕████████████████████████████████████████████████████████▏ 1.4 KB
pulling fcc5a6bec9da... 100% ▕████████████████████████████████████████████████████████▏ 7.7 KB
pulling a70ff7e570d9... 100% ▕████████████████████████████████████████████████████████▏ 6.0 KB
pulling 4f659a1e86d7... 100% ▕████████████████████████████████████████████████████████▏ 485 B
verifying sha256 digest
writing manifest
success
Error: error starting the external llama server: fork/exec /usr/lib/ollama/runners/cpu_avx2/ollama_llama_server: exec format error
### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
_No response_
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8067/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8067/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/134
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/134/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/134/comments
|
https://api.github.com/repos/ollama/ollama/issues/134/events
|
https://github.com/ollama/ollama/pull/134
| 1,813,267,553
|
PR_kwDOJ0Z1Ps5V9on7
| 134
|
Update images.go
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-20T06:34:19
| 2023-07-20T06:46:19
| 2023-07-20T06:46:02
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/134",
"html_url": "https://github.com/ollama/ollama/pull/134",
"diff_url": "https://github.com/ollama/ollama/pull/134.diff",
"patch_url": "https://github.com/ollama/ollama/pull/134.patch",
"merged_at": "2023-07-20T06:46:02"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/134/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/134/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3226
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3226/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3226/comments
|
https://api.github.com/repos/ollama/ollama/issues/3226/events
|
https://github.com/ollama/ollama/issues/3226
| 2,192,612,235
|
I_kwDOJ0Z1Ps6CsJuL
| 3,226
|
Cannot create Modelfile on Windows Shell
|
{
"login": "worras2019",
"id": 52268635,
"node_id": "MDQ6VXNlcjUyMjY4NjM1",
"avatar_url": "https://avatars.githubusercontent.com/u/52268635?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/worras2019",
"html_url": "https://github.com/worras2019",
"followers_url": "https://api.github.com/users/worras2019/followers",
"following_url": "https://api.github.com/users/worras2019/following{/other_user}",
"gists_url": "https://api.github.com/users/worras2019/gists{/gist_id}",
"starred_url": "https://api.github.com/users/worras2019/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/worras2019/subscriptions",
"organizations_url": "https://api.github.com/users/worras2019/orgs",
"repos_url": "https://api.github.com/users/worras2019/repos",
"events_url": "https://api.github.com/users/worras2019/events{/privacy}",
"received_events_url": "https://api.github.com/users/worras2019/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-03-18T15:57:03
| 2024-03-19T08:50:15
| 2024-03-19T08:50:15
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
The following is not working in WIndows Shell. Windows Shell doesn't recognize FROM.
--------------------------------------------------
FROM llama2
# set the temperature to 1 [higher is more creative, lower is more coherent]
PARAMETER temperature 1
# set the system message
SYSTEM """
You are Mario from Super Mario Bros. Answer as Mario, the assistant, only.
"""
### What did you expect to see?
Cannot create Modelfile
### Steps to reproduce
_No response_
### Are there any recent changes that introduced the issue?
_No response_
### OS
Windows
### Architecture
_No response_
### Platform
_No response_
### Ollama version
_No response_
### GPU
_No response_
### GPU info
_No response_
### CPU
_No response_
### Other software
_No response_
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3226/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3226/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/2788
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2788/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2788/comments
|
https://api.github.com/repos/ollama/ollama/issues/2788/events
|
https://github.com/ollama/ollama/issues/2788
| 2,157,663,506
|
I_kwDOJ0Z1Ps6Am1US
| 2,788
|
Bug: LLaVA 1.6 34b not respecting initial user prompt
|
{
"login": "mobilemike",
"id": 30590,
"node_id": "MDQ6VXNlcjMwNTkw",
"avatar_url": "https://avatars.githubusercontent.com/u/30590?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mobilemike",
"html_url": "https://github.com/mobilemike",
"followers_url": "https://api.github.com/users/mobilemike/followers",
"following_url": "https://api.github.com/users/mobilemike/following{/other_user}",
"gists_url": "https://api.github.com/users/mobilemike/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mobilemike/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mobilemike/subscriptions",
"organizations_url": "https://api.github.com/users/mobilemike/orgs",
"repos_url": "https://api.github.com/users/mobilemike/repos",
"events_url": "https://api.github.com/users/mobilemike/events{/privacy}",
"received_events_url": "https://api.github.com/users/mobilemike/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-27T21:47:53
| 2024-02-29T19:30:15
| 2024-02-29T19:30:15
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
M2 Max MBP 96GB RAM
Ollama 0.1.27
Compared against llama.cpp CLI @b11a93d (same as Ollama version)
Problem:
When using the Ollama CLI or API with an image, the initial text prompt isn't respected. Examples like the one on the [model page](https://ollama.com/library/llava) typically show prompts like "What is in this picture?". However, when changing the initial prompt to something like "Is this image of a llama?" or "How many animals are in this picture?" or even "Ignore the image and tell me the meaning of life", the output is typically a description of the image.
When using the llama.cpp CLI however, these prompts are followed as expected.
When using chat completions in Ollama, a followup question _does_ work properly, so after initially being ignored you can get the expected output on a second attempt.
My suspicion is that this behavior is largely unnoticed, as the default examples are asking for a description and one is being returned. However, this is masking the fact that text prompts used in conjunction with image prompts aren't being properly utilized. As an aside, LM Studio suffers from the same issue.
The below examples use the same image as the one base64 encoded in the above model page CLI example.
Ollama example:
```
❯ ollama run llava:34b-v1.6-q6_K
>>> /set parameter temperature 0.2
Set parameter 'temperature' to '0.2'
>>> How many animals are in this picture? /Users/mike/Downloads/llama.png
Added image '/Users/mike/Downloads/llama.png'
The image you've provided appears to be a cartoon or illustration of an
animal character. It looks like a cute, stylized depiction of a pig with a
happy expression and waving its hand as if saying hello or goodbye. The
art style is simplistic and playful, which is common in many modern
cartoons and emojis.
>>> How many animals are in this picture?
There is only one animal in this picture, which is the cute pig character.
```
llama.cpp example:
```
❯ ./llava-cli -m ../llm-models/cmp-nct/llava-1.6-gguf/ggml-yi-34b-f16-q_5_k.gguf --mmproj ../llm-models/cmp-nct/llava-1.6-gguf/mmproj-llava-34b-f16-q6_k.gguf --image ~/Downloads/llama.png --temp 0.2 -e -p '<|im_start|>system\n<|im_end|><|im_start|>user\n<image>\nHow many animals are in this picture?<|im_end|><|im_start|>assistant\n'
clip_model_load: model name: vit-large336-custom
clip_model_load: description: image encoder for LLaVA
clip_model_load: GGUF version: 3
clip_model_load: alignment: 32
clip_model_load: n_tensors: 378
clip_model_load: n_kv: 26
clip_model_load: ftype: q6_K
...
system_prompt: <|im_start|>system
<|im_end|><|im_start|>user
user_prompt:
How many animals are in this picture?<|im_end|><|im_start|>assistant
There is one animal in this picture, which appears to be a stylized drawing of a pig.
...
```
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2788/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2788/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4205
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4205/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4205/comments
|
https://api.github.com/repos/ollama/ollama/issues/4205/events
|
https://github.com/ollama/ollama/issues/4205
| 2,281,206,136
|
I_kwDOJ0Z1Ps6H-HF4
| 4,205
|
please support deepseek-ai/DeepSeek-V2-Chat
|
{
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/taozhiyuai/followers",
"following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}",
"gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}",
"starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions",
"organizations_url": "https://api.github.com/users/taozhiyuai/orgs",
"repos_url": "https://api.github.com/users/taozhiyuai/repos",
"events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}",
"received_events_url": "https://api.github.com/users/taozhiyuai/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-05-06T15:54:12
| 2024-05-08T06:12:39
| 2024-05-08T06:12:39
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
please support deepseek-ai/DeepSeek-V2-Chat
https://hf-mirror.com/deepseek-ai/DeepSeek-V2-Chat/tree/main
|
{
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/taozhiyuai/followers",
"following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}",
"gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}",
"starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions",
"organizations_url": "https://api.github.com/users/taozhiyuai/orgs",
"repos_url": "https://api.github.com/users/taozhiyuai/repos",
"events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}",
"received_events_url": "https://api.github.com/users/taozhiyuai/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4205/reactions",
"total_count": 3,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4205/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2058
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2058/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2058/comments
|
https://api.github.com/repos/ollama/ollama/issues/2058/events
|
https://github.com/ollama/ollama/issues/2058
| 2,089,071,525
|
I_kwDOJ0Z1Ps58hLOl
| 2,058
|
Prompt Eval Count is 1 when image is included in multimodal request
|
{
"login": "Dillon-Yun",
"id": 40805060,
"node_id": "MDQ6VXNlcjQwODA1MDYw",
"avatar_url": "https://avatars.githubusercontent.com/u/40805060?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Dillon-Yun",
"html_url": "https://github.com/Dillon-Yun",
"followers_url": "https://api.github.com/users/Dillon-Yun/followers",
"following_url": "https://api.github.com/users/Dillon-Yun/following{/other_user}",
"gists_url": "https://api.github.com/users/Dillon-Yun/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Dillon-Yun/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Dillon-Yun/subscriptions",
"organizations_url": "https://api.github.com/users/Dillon-Yun/orgs",
"repos_url": "https://api.github.com/users/Dillon-Yun/repos",
"events_url": "https://api.github.com/users/Dillon-Yun/events{/privacy}",
"received_events_url": "https://api.github.com/users/Dillon-Yun/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-01-18T21:24:42
| 2024-10-23T17:46:39
| 2024-10-23T17:46:38
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
When I hit the `/api/generate` endpoint for a multimodal model, such as bakllava, I am seeing that `prompt_eval_count == 1` when an image is included. Is this correct behavior? How can I retrieve the total amount of tokens in the request for a multimodal request?
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2058/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2058/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5880
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5880/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5880/comments
|
https://api.github.com/repos/ollama/ollama/issues/5880/events
|
https://github.com/ollama/ollama/issues/5880
| 2,425,575,429
|
I_kwDOJ0Z1Ps6Qk1gF
| 5,880
|
Make Ollama available on Chocolatey
|
{
"login": "b-a0",
"id": 25707742,
"node_id": "MDQ6VXNlcjI1NzA3NzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/25707742?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/b-a0",
"html_url": "https://github.com/b-a0",
"followers_url": "https://api.github.com/users/b-a0/followers",
"following_url": "https://api.github.com/users/b-a0/following{/other_user}",
"gists_url": "https://api.github.com/users/b-a0/gists{/gist_id}",
"starred_url": "https://api.github.com/users/b-a0/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/b-a0/subscriptions",
"organizations_url": "https://api.github.com/users/b-a0/orgs",
"repos_url": "https://api.github.com/users/b-a0/repos",
"events_url": "https://api.github.com/users/b-a0/events{/privacy}",
"received_events_url": "https://api.github.com/users/b-a0/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
}
] |
open
| false
| null |
[] | null | 2
| 2024-07-23T16:04:09
| 2024-09-05T19:52:37
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
The Windows installers works great, but requires that I manually download/update it from the ollama website. Would it be possible, probably when the Windows version has matured some more, to add ollama to [Chocolatey](https://community.chocolatey.org/)? That will probabably aid in making this an (even) easier to install package on Windows.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5880/reactions",
"total_count": 4,
"+1": 2,
"-1": 2,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5880/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/2377
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2377/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2377/comments
|
https://api.github.com/repos/ollama/ollama/issues/2377/events
|
https://github.com/ollama/ollama/pull/2377
| 2,121,583,534
|
PR_kwDOJ0Z1Ps5mL7sV
| 2,377
|
Bump llama.cpp to b2081
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-02-06T20:09:01
| 2024-02-07T20:04:41
| 2024-02-07T20:04:38
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2377",
"html_url": "https://github.com/ollama/ollama/pull/2377",
"diff_url": "https://github.com/ollama/ollama/pull/2377.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2377.patch",
"merged_at": "2024-02-07T20:04:38"
}
| null |
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2377/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2377/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3513
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3513/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3513/comments
|
https://api.github.com/repos/ollama/ollama/issues/3513/events
|
https://github.com/ollama/ollama/issues/3513
| 2,229,191,000
|
I_kwDOJ0Z1Ps6E3sFY
| 3,513
|
Stream: false is ignored in as of c863c6a96d01afda37b65f86eb14a1f04a3c7c47
|
{
"login": "eusebiu",
"id": 414589,
"node_id": "MDQ6VXNlcjQxNDU4OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eusebiu",
"html_url": "https://github.com/eusebiu",
"followers_url": "https://api.github.com/users/eusebiu/followers",
"following_url": "https://api.github.com/users/eusebiu/following{/other_user}",
"gists_url": "https://api.github.com/users/eusebiu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/eusebiu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eusebiu/subscriptions",
"organizations_url": "https://api.github.com/users/eusebiu/orgs",
"repos_url": "https://api.github.com/users/eusebiu/repos",
"events_url": "https://api.github.com/users/eusebiu/events{/privacy}",
"received_events_url": "https://api.github.com/users/eusebiu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-04-06T10:57:21
| 2024-04-13T07:22:27
| 2024-04-13T07:22:26
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Stream: false is ignored
### What did you expect to see?
When stream is set to false, the response should contain the full result.
### Steps to reproduce
Pull the latest code, compile, run (pull some model), run /api/generate with stream: false -> the result has no content (just some new lines) and "done: false".
### Are there any recent changes that introduced the issue?
I think this https://github.com/ollama/ollama/commit/c863c6a96d01afda37b65f86eb14a1f04a3c7c47 breaks it (Tested before and after this commit).
### OS
Windows
### Architecture
amd64
### Platform
_No response_
### Ollama version
latest
### GPU
Nvidia
### GPU info
Quadro RTX 5000 with Max-Q Design
### CPU
Intel
### Other software
_No response_
|
{
"login": "eusebiu",
"id": 414589,
"node_id": "MDQ6VXNlcjQxNDU4OQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eusebiu",
"html_url": "https://github.com/eusebiu",
"followers_url": "https://api.github.com/users/eusebiu/followers",
"following_url": "https://api.github.com/users/eusebiu/following{/other_user}",
"gists_url": "https://api.github.com/users/eusebiu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/eusebiu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eusebiu/subscriptions",
"organizations_url": "https://api.github.com/users/eusebiu/orgs",
"repos_url": "https://api.github.com/users/eusebiu/repos",
"events_url": "https://api.github.com/users/eusebiu/events{/privacy}",
"received_events_url": "https://api.github.com/users/eusebiu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3513/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3513/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8117
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8117/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8117/comments
|
https://api.github.com/repos/ollama/ollama/issues/8117/events
|
https://github.com/ollama/ollama/pull/8117
| 2,742,219,356
|
PR_kwDOJ0Z1Ps6FWP5c
| 8,117
|
Examples(privategpt): update outdated langchain dependencies and add UnstructuredRSTLoader
|
{
"login": "KaymeKaydex",
"id": 43151027,
"node_id": "MDQ6VXNlcjQzMTUxMDI3",
"avatar_url": "https://avatars.githubusercontent.com/u/43151027?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/KaymeKaydex",
"html_url": "https://github.com/KaymeKaydex",
"followers_url": "https://api.github.com/users/KaymeKaydex/followers",
"following_url": "https://api.github.com/users/KaymeKaydex/following{/other_user}",
"gists_url": "https://api.github.com/users/KaymeKaydex/gists{/gist_id}",
"starred_url": "https://api.github.com/users/KaymeKaydex/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/KaymeKaydex/subscriptions",
"organizations_url": "https://api.github.com/users/KaymeKaydex/orgs",
"repos_url": "https://api.github.com/users/KaymeKaydex/repos",
"events_url": "https://api.github.com/users/KaymeKaydex/events{/privacy}",
"received_events_url": "https://api.github.com/users/KaymeKaydex/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3
| 2024-12-16T12:12:05
| 2025-01-13T19:26:24
| 2025-01-13T19:26:24
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8117",
"html_url": "https://github.com/ollama/ollama/pull/8117",
"diff_url": "https://github.com/ollama/ollama/pull/8117.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8117.patch",
"merged_at": null
}
|
This commit updates the langchain dependencies to the latest stable versions and adds the UnstructuredRSTLoader for loading RST files.
- Update langchain dependencies to the latest stable versions.
- Add UnstructuredRSTLoader for loading RST files.
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8117/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8117/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3074
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3074/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3074/comments
|
https://api.github.com/repos/ollama/ollama/issues/3074/events
|
https://github.com/ollama/ollama/issues/3074
| 2,181,146,552
|
I_kwDOJ0Z1Ps6CAae4
| 3,074
|
CUDA error: an illegal memory access was encountered
|
{
"login": "lizhichao999",
"id": 34128722,
"node_id": "MDQ6VXNlcjM0MTI4NzIy",
"avatar_url": "https://avatars.githubusercontent.com/u/34128722?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lizhichao999",
"html_url": "https://github.com/lizhichao999",
"followers_url": "https://api.github.com/users/lizhichao999/followers",
"following_url": "https://api.github.com/users/lizhichao999/following{/other_user}",
"gists_url": "https://api.github.com/users/lizhichao999/gists{/gist_id}",
"starred_url": "https://api.github.com/users/lizhichao999/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lizhichao999/subscriptions",
"organizations_url": "https://api.github.com/users/lizhichao999/orgs",
"repos_url": "https://api.github.com/users/lizhichao999/repos",
"events_url": "https://api.github.com/users/lizhichao999/events{/privacy}",
"received_events_url": "https://api.github.com/users/lizhichao999/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 6
| 2024-03-12T09:35:45
| 2024-07-24T23:14:51
| 2024-07-24T23:14:51
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
ollama run llama2
When I executed the command: ollama run llama2, an error occurred related to the data information.

I only have a 128G GPU. Do I need more than 2 GPUs? I can run it on my local laptop with 2 GPUs without any errors.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3074/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3074/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7704
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7704/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7704/comments
|
https://api.github.com/repos/ollama/ollama/issues/7704/events
|
https://github.com/ollama/ollama/issues/7704
| 2,665,123,889
|
I_kwDOJ0Z1Ps6e2pAx
| 7,704
|
On Windows 11 pro, it does work to right click "restart to update"
|
{
"login": "jack-liew",
"id": 102201538,
"node_id": "U_kgDOBhd4wg",
"avatar_url": "https://avatars.githubusercontent.com/u/102201538?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jack-liew",
"html_url": "https://github.com/jack-liew",
"followers_url": "https://api.github.com/users/jack-liew/followers",
"following_url": "https://api.github.com/users/jack-liew/following{/other_user}",
"gists_url": "https://api.github.com/users/jack-liew/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jack-liew/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jack-liew/subscriptions",
"organizations_url": "https://api.github.com/users/jack-liew/orgs",
"repos_url": "https://api.github.com/users/jack-liew/repos",
"events_url": "https://api.github.com/users/jack-liew/events{/privacy}",
"received_events_url": "https://api.github.com/users/jack-liew/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
}
] |
closed
| false
| null |
[] | null | 1
| 2024-11-17T00:32:03
| 2024-11-18T22:39:53
| 2024-11-18T22:39:53
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
It does work to right click "restart to update".
Thanks fo great work.
----------------------
### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
0.4.1
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7704/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7704/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8642
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8642/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8642/comments
|
https://api.github.com/repos/ollama/ollama/issues/8642/events
|
https://github.com/ollama/ollama/pull/8642
| 2,816,895,956
|
PR_kwDOJ0Z1Ps6JSssP
| 8,642
|
Update README with DeepSeek model in the examples
|
{
"login": "harsha-s",
"id": 29120299,
"node_id": "MDQ6VXNlcjI5MTIwMjk5",
"avatar_url": "https://avatars.githubusercontent.com/u/29120299?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/harsha-s",
"html_url": "https://github.com/harsha-s",
"followers_url": "https://api.github.com/users/harsha-s/followers",
"following_url": "https://api.github.com/users/harsha-s/following{/other_user}",
"gists_url": "https://api.github.com/users/harsha-s/gists{/gist_id}",
"starred_url": "https://api.github.com/users/harsha-s/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/harsha-s/subscriptions",
"organizations_url": "https://api.github.com/users/harsha-s/orgs",
"repos_url": "https://api.github.com/users/harsha-s/repos",
"events_url": "https://api.github.com/users/harsha-s/events{/privacy}",
"received_events_url": "https://api.github.com/users/harsha-s/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 0
| 2025-01-28T23:10:31
| 2025-01-28T23:15:57
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8642",
"html_url": "https://github.com/ollama/ollama/pull/8642",
"diff_url": "https://github.com/ollama/ollama/pull/8642.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8642.patch",
"merged_at": null
}
|
Updated the README with the an example of DeepSeek model
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8642/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8642/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7306
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7306/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7306/comments
|
https://api.github.com/repos/ollama/ollama/issues/7306/events
|
https://github.com/ollama/ollama/pull/7306
| 2,603,930,359
|
PR_kwDOJ0Z1Ps5_Y0rC
| 7,306
|
integration: harden embedding test
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-10-22T00:15:14
| 2024-10-22T22:25:25
| 2024-10-22T22:25:22
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7306",
"html_url": "https://github.com/ollama/ollama/pull/7306",
"diff_url": "https://github.com/ollama/ollama/pull/7306.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7306.patch",
"merged_at": "2024-10-22T22:25:22"
}
|
Use cosine similarity to make the embeddings tests more robust
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7306/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7306/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1671
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1671/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1671/comments
|
https://api.github.com/repos/ollama/ollama/issues/1671/events
|
https://github.com/ollama/ollama/issues/1671
| 2,053,904,792
|
I_kwDOJ0Z1Ps56bBmY
| 1,671
|
Error: connect ECONNREFUSED 127.0.0.1:11434
|
{
"login": "LTtt456c",
"id": 53040718,
"node_id": "MDQ6VXNlcjUzMDQwNzE4",
"avatar_url": "https://avatars.githubusercontent.com/u/53040718?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/LTtt456c",
"html_url": "https://github.com/LTtt456c",
"followers_url": "https://api.github.com/users/LTtt456c/followers",
"following_url": "https://api.github.com/users/LTtt456c/following{/other_user}",
"gists_url": "https://api.github.com/users/LTtt456c/gists{/gist_id}",
"starred_url": "https://api.github.com/users/LTtt456c/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/LTtt456c/subscriptions",
"organizations_url": "https://api.github.com/users/LTtt456c/orgs",
"repos_url": "https://api.github.com/users/LTtt456c/repos",
"events_url": "https://api.github.com/users/LTtt456c/events{/privacy}",
"received_events_url": "https://api.github.com/users/LTtt456c/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 8
| 2023-12-22T13:12:25
| 2024-06-14T10:53:14
| 2023-12-29T13:49:09
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello everyone!
My ollama in My docker
docker Start ollama command is docker run -e OLLAMA_HOST=0.0.0.0:11434 -d -v ollama serve -p 11434:11434 --name ollama ollama/ollama
Then I in vscode open chatbot-ollama And then input npm run dev And then Report an error
↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓ Here is the error log ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓
PS G:\AI\chatbot-ollama> npm run dev
chatbot-ollama@0.1.0 dev
next dev
▲ Next.js 13.5.6
Local: http://localhost:3000/
✓ Ready in 2.9s
○ Compiling / ...
✓ Compiled / in 3.3s (1652 modules)
⚠ Fast Refresh had to perform a full reload. Read more: https://nextjs.org/docs/messages/fast-refresh-reload
✓ Compiled in 1699ms (1652 modules)
✓ Compiled in 519ms (1652 modules)
✓ Compiled /api/models in 245ms (68 modules)
[TypeError: fetch failed] {
cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] {
errno: -4078,
code: 'ECONNREFUSED',
syscall: 'connect',
address: '127.0.0.1',
port: 11434
}
}
✓ Compiled in 620ms (1720 modules)
[TypeError: fetch failed] {
cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] {
errno: -4078,
code: 'ECONNREFUSED',
syscall: 'connect',
address: '127.0.0.1',
port: 11434
}
}
[TypeError: fetch failed] {
cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] {
errno: -4078,
code: 'ECONNREFUSED',
syscall: 'connect',
address: '127.0.0.1',
port: 11434
}
}
[TypeError: fetch failed] {
cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] {
errno: -4078,
code: 'ECONNREFUSED',
syscall: 'connect',
address: '127.0.0.1',
port: 11434
}
}
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1671/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1671/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/756
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/756/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/756/comments
|
https://api.github.com/repos/ollama/ollama/issues/756/events
|
https://github.com/ollama/ollama/issues/756
| 1,938,044,665
|
I_kwDOJ0Z1Ps5zhDb5
| 756
|
Mistral - Failed To Load Model
|
{
"login": "mattdavenport",
"id": 1127393,
"node_id": "MDQ6VXNlcjExMjczOTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1127393?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mattdavenport",
"html_url": "https://github.com/mattdavenport",
"followers_url": "https://api.github.com/users/mattdavenport/followers",
"following_url": "https://api.github.com/users/mattdavenport/following{/other_user}",
"gists_url": "https://api.github.com/users/mattdavenport/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mattdavenport/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mattdavenport/subscriptions",
"organizations_url": "https://api.github.com/users/mattdavenport/orgs",
"repos_url": "https://api.github.com/users/mattdavenport/repos",
"events_url": "https://api.github.com/users/mattdavenport/events{/privacy}",
"received_events_url": "https://api.github.com/users/mattdavenport/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2023-10-11T15:08:01
| 2023-10-11T15:40:14
| 2023-10-11T15:10:58
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I'm running macOS (Ventura 13.0.1) 16in. M1 2021. I am able to run all of the llama2 models just fine, but the following occurs when attempting to run the mistral model:
```
~ % ollama pull mistral:latest
pulling manifest
pulling 6ae280299950... 100% |███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| (4.1/4.1 GB, 48 MB/s)
pulling fede2d8d6c1f... 100% |████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| (29/29 B, 194 kB/s)
pulling b96850d2e482... 100% |██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| (307/307 B, 1.4 MB/s)
verifying sha256 digest
writing manifest
success
~ % ollama run mistral:latest
>>> Hello
Error: failed to load model
```
If this is still a WIP please close this issue. The only other information I could find is the following log entries:
```
2023/10/11 10:25:47 images.go:1093: redirected to: https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/6a/6ae28029995007a3ee8d0b8556d50f3b59b831074cf19c84de87acf51fb54054/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20231011%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20231011T142546Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=386696851dae4763d830fc88c05381be653dab1e21243686e3180c01011644b6
2023/10/11 10:27:13 images.go:1061: success getting sha256:6ae28029995007a3ee8d0b8556d50f3b59b831074cf19c84de87acf51fb54054
2023/10/11 10:27:14 images.go:1093: redirected to: https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/fe/fede2d8d6c1f404b1db73b1cd26f7d5455ff2deeb737b5e2b339339dce2969d4/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20231011%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20231011T142714Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=4b1d208c4dcb6b20ae9727869c284c8ec0f77ee382ee975d96f50f1c358047e7
2023/10/11 10:27:14 images.go:1061: success getting sha256:fede2d8d6c1f404b1db73b1cd26f7d5455ff2deeb737b5e2b339339dce2969d4
2023/10/11 10:27:15 images.go:1093: redirected to: https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/b9/b96850d2e482b0d1af356eda4ac158af93e9b00e71363a9173d7b5480680bcf3/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20231011%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20231011T142715Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=a5cb6b9a865ca7264746ec67325b71ad80987c800bbeeadd43eef75a6e0363bc
2023/10/11 10:27:15 images.go:1061: success getting sha256:b96850d2e482b0d1af356eda4ac158af93e9b00e71363a9173d7b5480680bcf3
[GIN] 2023/10/11 - 10:27:18 | 200 | 1m33s | 127.0.0.1 | POST "/api/pull"
llama.cpp: loading model from /Users/mattdavenport/.ollama/models/blobs/sha256:6ae28029995007a3ee8d0b8556d50f3b59b831074cf19c84de87acf51fb54054
error loading model: unknown (magic, version) combination: 46554747, 00000002; is this really a GGML file?
llama_load_model_from_file: failed to load model
[GIN] 2023/10/11 - 11:04:20 | 500 | 3.950083ms | 127.0.0.1 | POST "/api/generate"
```
Please let me know if I can provide any additional information here to help debug. Thanks!
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/756/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/756/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2363
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2363/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2363/comments
|
https://api.github.com/repos/ollama/ollama/issues/2363/events
|
https://github.com/ollama/ollama/pull/2363
| 2,119,594,984
|
PR_kwDOJ0Z1Ps5mFH5O
| 2,363
|
Move hub auth out to new package
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-02-05T22:39:49
| 2024-02-15T17:42:21
| 2024-02-15T17:42:17
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2363",
"html_url": "https://github.com/ollama/ollama/pull/2363",
"diff_url": "https://github.com/ollama/ollama/pull/2363.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2363.patch",
"merged_at": null
}
| null |
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2363/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2363/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2514
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2514/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2514/comments
|
https://api.github.com/repos/ollama/ollama/issues/2514/events
|
https://github.com/ollama/ollama/issues/2514
| 2,136,760,146
|
I_kwDOJ0Z1Ps5_XF9S
| 2,514
|
Support for safetensors
|
{
"login": "gracehubai",
"id": 159539863,
"node_id": "U_kgDOCYJilw",
"avatar_url": "https://avatars.githubusercontent.com/u/159539863?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gracehubai",
"html_url": "https://github.com/gracehubai",
"followers_url": "https://api.github.com/users/gracehubai/followers",
"following_url": "https://api.github.com/users/gracehubai/following{/other_user}",
"gists_url": "https://api.github.com/users/gracehubai/gists{/gist_id}",
"starred_url": "https://api.github.com/users/gracehubai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gracehubai/subscriptions",
"organizations_url": "https://api.github.com/users/gracehubai/orgs",
"repos_url": "https://api.github.com/users/gracehubai/repos",
"events_url": "https://api.github.com/users/gracehubai/events{/privacy}",
"received_events_url": "https://api.github.com/users/gracehubai/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2024-02-15T15:02:55
| 2024-02-21T05:38:43
| 2024-02-18T06:49:05
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Do we already support inferencing safetensors?
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2514/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2514/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2214
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2214/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2214/comments
|
https://api.github.com/repos/ollama/ollama/issues/2214/events
|
https://github.com/ollama/ollama/pull/2214
| 2,102,794,214
|
PR_kwDOJ0Z1Ps5lMKh3
| 2,214
|
Detect lack of AVX and fallback to CPU mode
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-01-26T19:42:11
| 2024-01-26T20:06:47
| 2024-01-26T20:06:44
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2214",
"html_url": "https://github.com/ollama/ollama/pull/2214",
"diff_url": "https://github.com/ollama/ollama/pull/2214.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2214.patch",
"merged_at": "2024-01-26T20:06:44"
}
|
We build the GPU libraries with AVX enabled to ensure that if not all layers fit on the GPU we get better performance in a mixed mode. If the user is using a virtualization/emulation system that lacks AVX this used to result in an illegal instruction error and crash before this fix. Now we will report a warning in the server log, and just use CPU mode to ensure we don't crash.
This should mitigate #2187.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2214/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2214/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8431
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8431/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8431/comments
|
https://api.github.com/repos/ollama/ollama/issues/8431/events
|
https://github.com/ollama/ollama/issues/8431
| 2,788,732,623
|
I_kwDOJ0Z1Ps6mOK7P
| 8,431
|
Embedding with bge-large:latest fails with "GGML_ASSERT(i01 >= 0 && i01 < ne01) failed"
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3
| 2025-01-15T03:07:10
| 2025-01-15T06:35:56
| null |
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
> @rick-github Thanks,I tried the "options":{"num_ctx":512} ,It works well! 💯
> but the logs always shows that
>
> 
_Originally posted by @Justin-12138 in [#7441](https://github.com/ollama/ollama/issues/7441#issuecomment-2591480334)_
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8431/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8431/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/1128
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1128/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1128/comments
|
https://api.github.com/repos/ollama/ollama/issues/1128/events
|
https://github.com/ollama/ollama/pull/1128
| 1,993,443,855
|
PR_kwDOJ0Z1Ps5fcrpD
| 1,128
|
FAQ: answer a few faq questions
|
{
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.github.com/users/technovangelist/followers",
"following_url": "https://api.github.com/users/technovangelist/following{/other_user}",
"gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}",
"starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions",
"organizations_url": "https://api.github.com/users/technovangelist/orgs",
"repos_url": "https://api.github.com/users/technovangelist/repos",
"events_url": "https://api.github.com/users/technovangelist/events{/privacy}",
"received_events_url": "https://api.github.com/users/technovangelist/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-11-14T19:43:51
| 2023-11-15T23:05:14
| 2023-11-15T23:05:13
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1128",
"html_url": "https://github.com/ollama/ollama/pull/1128",
"diff_url": "https://github.com/ollama/ollama/pull/1128.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1128.patch",
"merged_at": "2023-11-15T23:05:13"
}
| null |
{
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/users/mchiang0610/followers",
"following_url": "https://api.github.com/users/mchiang0610/following{/other_user}",
"gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions",
"organizations_url": "https://api.github.com/users/mchiang0610/orgs",
"repos_url": "https://api.github.com/users/mchiang0610/repos",
"events_url": "https://api.github.com/users/mchiang0610/events{/privacy}",
"received_events_url": "https://api.github.com/users/mchiang0610/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1128/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1128/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4247
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4247/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4247/comments
|
https://api.github.com/repos/ollama/ollama/issues/4247/events
|
https://github.com/ollama/ollama/pull/4247
| 2,284,495,403
|
PR_kwDOJ0Z1Ps5u0qMP
| 4,247
|
skip hidden files in list models handler
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-05-08T01:34:13
| 2024-05-08T02:01:46
| 2024-05-08T02:01:45
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4247",
"html_url": "https://github.com/ollama/ollama/pull/4247",
"diff_url": "https://github.com/ollama/ollama/pull/4247.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4247.patch",
"merged_at": "2024-05-08T02:01:45"
}
|
Hidden files on MacOS (ex: `.DS_Store`) cause the list command to file when not skipped:
```
❯ ollama ls
Error: unqualified name:
```
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4247/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4247/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4025
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4025/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4025/comments
|
https://api.github.com/repos/ollama/ollama/issues/4025/events
|
https://github.com/ollama/ollama/pull/4025
| 2,269,092,269
|
PR_kwDOJ0Z1Ps5uBD92
| 4,025
|
Update linux.md, suggest containerized install
|
{
"login": "ericcurtin",
"id": 1694275,
"node_id": "MDQ6VXNlcjE2OTQyNzU=",
"avatar_url": "https://avatars.githubusercontent.com/u/1694275?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ericcurtin",
"html_url": "https://github.com/ericcurtin",
"followers_url": "https://api.github.com/users/ericcurtin/followers",
"following_url": "https://api.github.com/users/ericcurtin/following{/other_user}",
"gists_url": "https://api.github.com/users/ericcurtin/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ericcurtin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ericcurtin/subscriptions",
"organizations_url": "https://api.github.com/users/ericcurtin/orgs",
"repos_url": "https://api.github.com/users/ericcurtin/repos",
"events_url": "https://api.github.com/users/ericcurtin/events{/privacy}",
"received_events_url": "https://api.github.com/users/ericcurtin/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 2
| 2024-04-29T13:43:19
| 2024-05-06T22:54:56
| null |
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4025",
"html_url": "https://github.com/ollama/ollama/pull/4025",
"diff_url": "https://github.com/ollama/ollama/pull/4025.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4025.patch",
"merged_at": null
}
|
It's the simplest install process, once you run the one liner, everything just works.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4025/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4025/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5860
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5860/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5860/comments
|
https://api.github.com/repos/ollama/ollama/issues/5860/events
|
https://github.com/ollama/ollama/pull/5860
| 2,423,755,415
|
PR_kwDOJ0Z1Ps52I6Yu
| 5,860
|
auth: update auth
|
{
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/joshyan1/followers",
"following_url": "https://api.github.com/users/joshyan1/following{/other_user}",
"gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions",
"organizations_url": "https://api.github.com/users/joshyan1/orgs",
"repos_url": "https://api.github.com/users/joshyan1/repos",
"events_url": "https://api.github.com/users/joshyan1/events{/privacy}",
"received_events_url": "https://api.github.com/users/joshyan1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-07-22T20:55:20
| 2024-09-04T13:43:34
| 2024-09-04T13:43:33
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5860",
"html_url": "https://github.com/ollama/ollama/pull/5860",
"diff_url": "https://github.com/ollama/ollama/pull/5860.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5860.patch",
"merged_at": null
}
|
pre cursor to setting up tests for https://github.com/ollama/ollama/pull/5441
moved initialize keypair to auth
use initialize keypair implicitly when reading pubkey
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5860/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5860/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/511
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/511/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/511/comments
|
https://api.github.com/repos/ollama/ollama/issues/511/events
|
https://github.com/ollama/ollama/issues/511
| 1,891,545,545
|
I_kwDOJ0Z1Ps5wvrHJ
| 511
|
Support InternLM
|
{
"login": "vansinhu",
"id": 60632596,
"node_id": "MDQ6VXNlcjYwNjMyNTk2",
"avatar_url": "https://avatars.githubusercontent.com/u/60632596?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vansinhu",
"html_url": "https://github.com/vansinhu",
"followers_url": "https://api.github.com/users/vansinhu/followers",
"following_url": "https://api.github.com/users/vansinhu/following{/other_user}",
"gists_url": "https://api.github.com/users/vansinhu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/vansinhu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vansinhu/subscriptions",
"organizations_url": "https://api.github.com/users/vansinhu/orgs",
"repos_url": "https://api.github.com/users/vansinhu/repos",
"events_url": "https://api.github.com/users/vansinhu/events{/privacy}",
"received_events_url": "https://api.github.com/users/vansinhu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 6
| 2023-09-12T02:09:05
| 2024-06-24T08:28:12
| 2024-03-11T19:32:53
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Dear ollama developer,
Greetings! I am vansinhu, a community developer and volunteer at InternLM. [InternLM](https://github.com/InternLM/InternLM) is a large language model similar to llama2, and we look forward to InternLM being supported in ollama. If there are any challenges or inquiries regarding support for InternLM, please feel free to join our Discord discussion at https://discord.gg/gF9ezcmtM3.
Best regards,
vansinhu
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/511/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/511/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8383
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8383/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8383/comments
|
https://api.github.com/repos/ollama/ollama/issues/8383/events
|
https://github.com/ollama/ollama/pull/8383
| 2,781,700,019
|
PR_kwDOJ0Z1Ps6Hah72
| 8,383
|
fix: run container as non-root user
|
{
"login": "chaturvedi-kna",
"id": 63336082,
"node_id": "MDQ6VXNlcjYzMzM2MDgy",
"avatar_url": "https://avatars.githubusercontent.com/u/63336082?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chaturvedi-kna",
"html_url": "https://github.com/chaturvedi-kna",
"followers_url": "https://api.github.com/users/chaturvedi-kna/followers",
"following_url": "https://api.github.com/users/chaturvedi-kna/following{/other_user}",
"gists_url": "https://api.github.com/users/chaturvedi-kna/gists{/gist_id}",
"starred_url": "https://api.github.com/users/chaturvedi-kna/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/chaturvedi-kna/subscriptions",
"organizations_url": "https://api.github.com/users/chaturvedi-kna/orgs",
"repos_url": "https://api.github.com/users/chaturvedi-kna/repos",
"events_url": "https://api.github.com/users/chaturvedi-kna/events{/privacy}",
"received_events_url": "https://api.github.com/users/chaturvedi-kna/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 1
| 2025-01-11T08:28:55
| 2025-01-19T02:54:33
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8383",
"html_url": "https://github.com/ollama/ollama/pull/8383",
"diff_url": "https://github.com/ollama/ollama/pull/8383.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8383.patch",
"merged_at": null
}
|
**Enable non-root user Container Deployment Support**
**Description**
This PR adds support for running ollama in rootless containers by properly setting up permissions for the `.ollama` directory. This change enables ollama to run in environments that enforce non-root execution such as:
* OpenShift
* Kubernetes with strict security contexts
* Containers with user namespace remapping
**Changes**
* Added directory creation and permission setup in the final stage of Dockerfile
* Set container to run as non-root user (UID 1001)
* Configured group permissions to support random GID assignment
**Testing Done**
* Built image with podman
* Deployed on OpenShift
* Verified container starts successfully with non-root user
* Confirmed `.ollama` directory is properly accessible
**Previous Behavior**
The container would fail to start in rootless environments with the error:
```
Couldn't find '/.ollama/id_ed25519'. Generating new private key.
Error: could not create directory mkdir /.ollama: permission denied
```
**New Behavior**
* Container starts successfully as non-root user
* `.ollama` directory is created with proper permissions
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8383/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8383/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5306
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5306/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5306/comments
|
https://api.github.com/repos/ollama/ollama/issues/5306/events
|
https://github.com/ollama/ollama/pull/5306
| 2,375,887,192
|
PR_kwDOJ0Z1Ps5zqhZW
| 5,306
|
Do not reinstall the CLI tools if they are already installed on macOS
|
{
"login": "seanchristians",
"id": 25487785,
"node_id": "MDQ6VXNlcjI1NDg3Nzg1",
"avatar_url": "https://avatars.githubusercontent.com/u/25487785?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/seanchristians",
"html_url": "https://github.com/seanchristians",
"followers_url": "https://api.github.com/users/seanchristians/followers",
"following_url": "https://api.github.com/users/seanchristians/following{/other_user}",
"gists_url": "https://api.github.com/users/seanchristians/gists{/gist_id}",
"starred_url": "https://api.github.com/users/seanchristians/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/seanchristians/subscriptions",
"organizations_url": "https://api.github.com/users/seanchristians/orgs",
"repos_url": "https://api.github.com/users/seanchristians/repos",
"events_url": "https://api.github.com/users/seanchristians/events{/privacy}",
"received_events_url": "https://api.github.com/users/seanchristians/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-06-26T17:17:30
| 2024-09-06T17:49:40
| 2024-09-05T04:41:02
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5306",
"html_url": "https://github.com/ollama/ollama/pull/5306",
"diff_url": "https://github.com/ollama/ollama/pull/5306.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5306.patch",
"merged_at": null
}
|
Check `installed()` inside of the `install` function before running the cli setup commands. Fix #5305.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5306/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5306/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6165
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6165/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6165/comments
|
https://api.github.com/repos/ollama/ollama/issues/6165/events
|
https://github.com/ollama/ollama/issues/6165
| 2,447,319,803
|
I_kwDOJ0Z1Ps6R3yL7
| 6,165
|
No devices found using AMD gpus
|
{
"login": "cinglish",
"id": 40436106,
"node_id": "MDQ6VXNlcjQwNDM2MTA2",
"avatar_url": "https://avatars.githubusercontent.com/u/40436106?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/cinglish",
"html_url": "https://github.com/cinglish",
"followers_url": "https://api.github.com/users/cinglish/followers",
"following_url": "https://api.github.com/users/cinglish/following{/other_user}",
"gists_url": "https://api.github.com/users/cinglish/gists{/gist_id}",
"starred_url": "https://api.github.com/users/cinglish/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/cinglish/subscriptions",
"organizations_url": "https://api.github.com/users/cinglish/orgs",
"repos_url": "https://api.github.com/users/cinglish/repos",
"events_url": "https://api.github.com/users/cinglish/events{/privacy}",
"received_events_url": "https://api.github.com/users/cinglish/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg",
"url": "https://api.github.com/repos/ollama/ollama/labels/linux",
"name": "linux",
"color": "516E70",
"default": false,
"description": ""
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
},
{
"id": 6677677816,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgVG-A",
"url": "https://api.github.com/repos/ollama/ollama/labels/docker",
"name": "docker",
"color": "0052CC",
"default": false,
"description": "Issues relating to using ollama in containers"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 2
| 2024-08-04T22:23:38
| 2024-09-05T19:03:53
| 2024-09-05T19:03:53
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Getting the following error when loading models with AMD gpus (Instinct MI60s):
```
rocBLAS error: Could not initialize Tensile host: No devices found
```
I have 4 devices allocated to the container and it seems to still discover them at startup:
```
time=2024-08-04T22:17:48.256Z level=INFO source=gpu.go:205 msg="looking for compatible GPUs"
time=2024-08-04T22:17:48.264Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=0 gpu_type=gfx906
time=2024-08-04T22:17:48.264Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=1
time=2024-08-04T22:17:48.264Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=2
time=2024-08-04T22:17:48.264Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=3
time=2024-08-04T22:17:48.265Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=4
time=2024-08-04T22:17:48.265Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=5 gpu_type=gfx906
time=2024-08-04T22:17:48.266Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=6 gpu_type=gfx906
time=2024-08-04T22:17:48.268Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=7 gpu_type=gfx906
```
It is working fine with the `0.3.1-rocm` image, but seeing this behavior in the newest versions.
### OS
Docker
### GPU
AMD
### CPU
Intel
### Ollama version
0.3.2
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6165/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6165/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2989
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2989/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2989/comments
|
https://api.github.com/repos/ollama/ollama/issues/2989/events
|
https://github.com/ollama/ollama/pull/2989
| 2,174,611,866
|
PR_kwDOJ0Z1Ps5pAaFv
| 2,989
|
CI updates
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-03-07T19:26:30
| 2024-03-09T21:20:27
| 2024-03-09T21:20:24
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2989",
"html_url": "https://github.com/ollama/ollama/pull/2989",
"diff_url": "https://github.com/ollama/ollama/pull/2989.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2989.patch",
"merged_at": null
}
|
Not ready for review yet...
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2989/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2989/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/476
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/476/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/476/comments
|
https://api.github.com/repos/ollama/ollama/issues/476/events
|
https://github.com/ollama/ollama/pull/476
| 1,884,716,024
|
PR_kwDOJ0Z1Ps5ZuHR7
| 476
|
tighten up the error string for `ollama show` flags
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-09-06T20:37:08
| 2023-09-06T20:38:50
| 2023-09-06T20:38:49
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/476",
"html_url": "https://github.com/ollama/ollama/pull/476",
"diff_url": "https://github.com/ollama/ollama/pull/476.diff",
"patch_url": "https://github.com/ollama/ollama/pull/476.patch",
"merged_at": "2023-09-06T20:38:49"
}
| null |
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/476/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/476/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3018
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3018/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3018/comments
|
https://api.github.com/repos/ollama/ollama/issues/3018/events
|
https://github.com/ollama/ollama/pull/3018
| 2,177,073,503
|
PR_kwDOJ0Z1Ps5pIy3-
| 3,018
|
Add allowed host middleware and remove `workDir` middleware
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-03-09T06:14:34
| 2024-03-09T06:23:48
| 2024-03-09T06:23:47
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3018",
"html_url": "https://github.com/ollama/ollama/pull/3018",
"diff_url": "https://github.com/ollama/ollama/pull/3018.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3018.patch",
"merged_at": "2024-03-09T06:23:47"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3018/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3018/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/638
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/638/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/638/comments
|
https://api.github.com/repos/ollama/ollama/issues/638/events
|
https://github.com/ollama/ollama/pull/638
| 1,918,206,635
|
PR_kwDOJ0Z1Ps5beytn
| 638
|
add cuda to windows generation
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-09-28T20:15:47
| 2023-10-20T16:44:24
| 2023-09-29T16:53:14
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/638",
"html_url": "https://github.com/ollama/ollama/pull/638",
"diff_url": "https://github.com/ollama/ollama/pull/638.diff",
"patch_url": "https://github.com/ollama/ollama/pull/638.patch",
"merged_at": null
}
|
Still a bunch of stuff to do for packaging this for distribution, but I tested building and running this locally and it works along with #637 so may as well get it in.
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/638/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/638/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3616
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3616/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3616/comments
|
https://api.github.com/repos/ollama/ollama/issues/3616/events
|
https://github.com/ollama/ollama/issues/3616
| 2,240,686,081
|
I_kwDOJ0Z1Ps6FjigB
| 3,616
|
Richer grammars
|
{
"login": "tezlm",
"id": 56274831,
"node_id": "MDQ6VXNlcjU2Mjc0ODMx",
"avatar_url": "https://avatars.githubusercontent.com/u/56274831?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tezlm",
"html_url": "https://github.com/tezlm",
"followers_url": "https://api.github.com/users/tezlm/followers",
"following_url": "https://api.github.com/users/tezlm/following{/other_user}",
"gists_url": "https://api.github.com/users/tezlm/gists{/gist_id}",
"starred_url": "https://api.github.com/users/tezlm/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/tezlm/subscriptions",
"organizations_url": "https://api.github.com/users/tezlm/orgs",
"repos_url": "https://api.github.com/users/tezlm/repos",
"events_url": "https://api.github.com/users/tezlm/events{/privacy}",
"received_events_url": "https://api.github.com/users/tezlm/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 9
| 2024-04-12T18:23:12
| 2024-12-05T00:37:28
| 2024-12-05T00:37:28
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What are you trying to do?
Being able to specify grammars is great, but it seems a bit underutilized at the moment. This is mostly a thought dump on how it could be improved from experimentation...
### How should we solve this?
- Using llama.cpp grammar directly would be pretty powerful and nice to have
- Specifying jsonschema for json. Llama.cpp json is usually forced into a specific key order and ollama json isn't schema'd at all
- Changing the format on the fly is useful, but I think it would be nice to have a way to specify a grammar in the Modelfile.
### What is the impact of not solving this?
Not having either of the first two ideas is annoying, since there's no way to guarantee that a model generates a response in a format I want. The third idea allows one to make a "llm api", where a model generates a specific response every time (imagine bundling a "summary llm" that always responds with `{"summary":"..."}` as a Modelfile.)
### Anything else?
_No response_
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3616/reactions",
"total_count": 9,
"+1": 9,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3616/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8624
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8624/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8624/comments
|
https://api.github.com/repos/ollama/ollama/issues/8624/events
|
https://github.com/ollama/ollama/issues/8624
| 2,814,669,000
|
I_kwDOJ0Z1Ps6nxHDI
| 8,624
|
Deepseek 80% size reduction
|
{
"login": "gileneusz",
"id": 34601970,
"node_id": "MDQ6VXNlcjM0NjAxOTcw",
"avatar_url": "https://avatars.githubusercontent.com/u/34601970?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gileneusz",
"html_url": "https://github.com/gileneusz",
"followers_url": "https://api.github.com/users/gileneusz/followers",
"following_url": "https://api.github.com/users/gileneusz/following{/other_user}",
"gists_url": "https://api.github.com/users/gileneusz/gists{/gist_id}",
"starred_url": "https://api.github.com/users/gileneusz/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gileneusz/subscriptions",
"organizations_url": "https://api.github.com/users/gileneusz/orgs",
"repos_url": "https://api.github.com/users/gileneusz/repos",
"events_url": "https://api.github.com/users/gileneusz/events{/privacy}",
"received_events_url": "https://api.github.com/users/gileneusz/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
open
| false
| null |
[] | null | 1
| 2025-01-28T05:27:54
| 2025-01-28T22:28:24
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
New quants done by unsloth.ai:
| MoE Bits | Disk Size | Type | Quality | Link | Down_proj |
|-----------|-----------|----------|---------|------------------------------------------------------------------------------------------------------------------------------------------------|----------------|
| 1.58-bit | 131GB | IQ1_S | Fair | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-IQ1_S) | 2.06/1.56bit |
| 1.73-bit | 158GB | IQ1_M | Good | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-IQ1_M) | 2.06bit |
| 2.22-bit | 183GB | IQ2_XXS | Better | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-IQ2_XXS) | 2.5/2.06bit |
| 2.51-bit | 212GB | Q2_K_XL | Best | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-Q2_K_XL) | 3.5/2.5bit |
please consider adding them
https://unsloth.ai/blog/deepseekr1-dynamic
thanks!
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8624/reactions",
"total_count": 8,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 8,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8624/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/59
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/59/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/59/comments
|
https://api.github.com/repos/ollama/ollama/issues/59/events
|
https://github.com/ollama/ollama/issues/59
| 1,795,166,966
|
I_kwDOJ0Z1Ps5rABL2
| 59
|
generate pauses after about 50 tokens
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 2
| 2023-07-08T23:42:53
| 2023-07-13T02:15:17
| 2023-07-13T02:15:17
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Generation will get paused after about 50 tokens being provided
```
% ollama run orca
>>> Write a review of the restaurant "five guys"
As an AI assistant, I cannot write a biased or subjective review, but I can provide you with some general information about the restaurant "Five Guys". Five Guys is an American fast-food chain that primarily serves hamburgers, fries, <pause here>
```
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/59/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/59/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5278
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5278/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5278/comments
|
https://api.github.com/repos/ollama/ollama/issues/5278/events
|
https://github.com/ollama/ollama/issues/5278
| 2,373,277,484
|
I_kwDOJ0Z1Ps6NdVcs
| 5,278
|
Is it possible to start llama server through dynamic dependency library?
|
{
"login": "leeyiding",
"id": 49098278,
"node_id": "MDQ6VXNlcjQ5MDk4Mjc4",
"avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leeyiding",
"html_url": "https://github.com/leeyiding",
"followers_url": "https://api.github.com/users/leeyiding/followers",
"following_url": "https://api.github.com/users/leeyiding/following{/other_user}",
"gists_url": "https://api.github.com/users/leeyiding/gists{/gist_id}",
"starred_url": "https://api.github.com/users/leeyiding/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/leeyiding/subscriptions",
"organizations_url": "https://api.github.com/users/leeyiding/orgs",
"repos_url": "https://api.github.com/users/leeyiding/repos",
"events_url": "https://api.github.com/users/leeyiding/events{/privacy}",
"received_events_url": "https://api.github.com/users/leeyiding/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-06-25T17:43:21
| 2024-07-03T02:36:33
| 2024-07-03T02:36:33
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi, I'm trying to run Ollama in [Nanos Unikernel](https://github.com/nanovms/nanos), Unikernel a single-process operating system.
I found that in https://github.com/ollama/ollama/commit/58d95cc9bd446a8209e7388a96c70367cbafd653, the running mode of llama server changed from loading dynamic dependency libraries to running through subprocess. Therefore, before version 0.1.32, I could run Ollama normally in Nanos, but it could not run normally in later versions.
I would like to ask whether it is possible to continue to provide the method of loading llama server as a dynamic dependency library and avoid using subprocess. Perhaps the two methods can coexist and users can choose by themselves.
Looking forward to your reply, thank you very much.
|
{
"login": "leeyiding",
"id": 49098278,
"node_id": "MDQ6VXNlcjQ5MDk4Mjc4",
"avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leeyiding",
"html_url": "https://github.com/leeyiding",
"followers_url": "https://api.github.com/users/leeyiding/followers",
"following_url": "https://api.github.com/users/leeyiding/following{/other_user}",
"gists_url": "https://api.github.com/users/leeyiding/gists{/gist_id}",
"starred_url": "https://api.github.com/users/leeyiding/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/leeyiding/subscriptions",
"organizations_url": "https://api.github.com/users/leeyiding/orgs",
"repos_url": "https://api.github.com/users/leeyiding/repos",
"events_url": "https://api.github.com/users/leeyiding/events{/privacy}",
"received_events_url": "https://api.github.com/users/leeyiding/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5278/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5278/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1579
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1579/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1579/comments
|
https://api.github.com/repos/ollama/ollama/issues/1579/events
|
https://github.com/ollama/ollama/issues/1579
| 2,046,156,422
|
I_kwDOJ0Z1Ps559d6G
| 1,579
|
Error while running ollama locally.
|
{
"login": "nehalmathew1996",
"id": 82964954,
"node_id": "MDQ6VXNlcjgyOTY0OTU0",
"avatar_url": "https://avatars.githubusercontent.com/u/82964954?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nehalmathew1996",
"html_url": "https://github.com/nehalmathew1996",
"followers_url": "https://api.github.com/users/nehalmathew1996/followers",
"following_url": "https://api.github.com/users/nehalmathew1996/following{/other_user}",
"gists_url": "https://api.github.com/users/nehalmathew1996/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nehalmathew1996/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nehalmathew1996/subscriptions",
"organizations_url": "https://api.github.com/users/nehalmathew1996/orgs",
"repos_url": "https://api.github.com/users/nehalmathew1996/repos",
"events_url": "https://api.github.com/users/nehalmathew1996/events{/privacy}",
"received_events_url": "https://api.github.com/users/nehalmathew1996/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 11
| 2023-12-18T09:28:17
| 2024-04-05T21:14:06
| 2024-02-01T23:19:46
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
ConnectionError: HTTPConnectionPool(host='localhost', port=11434): Max retries exceeded with url: /api/generate/ (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x000002298AE1EF50>: Failed to establish a new connection: [WinError 10061] No connection could be made because the target machine actively refused it'))
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1579/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1579/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7394
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7394/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7394/comments
|
https://api.github.com/repos/ollama/ollama/issues/7394/events
|
https://github.com/ollama/ollama/issues/7394
| 2,617,770,363
|
I_kwDOJ0Z1Ps6cCAF7
| 7,394
|
The Open WebUI generate unsense text, but the cli terminal can chat normally with the same hosted model
|
{
"login": "QiuJYWX",
"id": 68810978,
"node_id": "MDQ6VXNlcjY4ODEwOTc4",
"avatar_url": "https://avatars.githubusercontent.com/u/68810978?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/QiuJYWX",
"html_url": "https://github.com/QiuJYWX",
"followers_url": "https://api.github.com/users/QiuJYWX/followers",
"following_url": "https://api.github.com/users/QiuJYWX/following{/other_user}",
"gists_url": "https://api.github.com/users/QiuJYWX/gists{/gist_id}",
"starred_url": "https://api.github.com/users/QiuJYWX/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/QiuJYWX/subscriptions",
"organizations_url": "https://api.github.com/users/QiuJYWX/orgs",
"repos_url": "https://api.github.com/users/QiuJYWX/repos",
"events_url": "https://api.github.com/users/QiuJYWX/events{/privacy}",
"received_events_url": "https://api.github.com/users/QiuJYWX/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] |
closed
| false
| null |
[] | null | 9
| 2024-10-28T09:06:43
| 2024-11-13T22:12:46
| 2024-11-13T22:12:46
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
The results generated by cli terminal:

The results generated by open webui:


### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.14
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7394/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7394/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8316
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8316/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8316/comments
|
https://api.github.com/repos/ollama/ollama/issues/8316/events
|
https://github.com/ollama/ollama/issues/8316
| 2,770,077,246
|
I_kwDOJ0Z1Ps6lHAY-
| 8,316
|
Feature Request: Limit Ollama to use only GPU RAM (VRAM), not system RAM
|
{
"login": "mlibre",
"id": 8473036,
"node_id": "MDQ6VXNlcjg0NzMwMzY=",
"avatar_url": "https://avatars.githubusercontent.com/u/8473036?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mlibre",
"html_url": "https://github.com/mlibre",
"followers_url": "https://api.github.com/users/mlibre/followers",
"following_url": "https://api.github.com/users/mlibre/following{/other_user}",
"gists_url": "https://api.github.com/users/mlibre/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mlibre/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mlibre/subscriptions",
"organizations_url": "https://api.github.com/users/mlibre/orgs",
"repos_url": "https://api.github.com/users/mlibre/repos",
"events_url": "https://api.github.com/users/mlibre/events{/privacy}",
"received_events_url": "https://api.github.com/users/mlibre/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 4
| 2025-01-06T08:02:41
| 2025-01-24T09:45:19
| 2025-01-24T09:45:19
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### Feature Request
**Description**:
I would like to request a feature that allows Ollama to be configured to use only GPU RAM, without utilizing CPU or system RAM. This would be beneficial for systems with limited RAM resources and high GPU RAM (VRAM) memory availability.
**Benefits**:
- Optimized performance on GPU-heavy systems
- Reduced CPU load and system RAM usage
- Enhanced efficiency for specific workloads
**Example Use Case**:
In scenarios where the system has multiple GPUs with high memory capacity but limited CPU and system RAM, this feature would enable Ollama to leverage the full potential of the GPUs without being constrained by CPU or system RAM limitations.
**Environment Variables/Configuration**:
- An environment variable or configuration setting to enforce GPU-only memory usage.
- Example: `OLLAMA_USE_GPU_ONLY=1`
**Related Files and Code**:
- Documentation: [docs/gpu.md](https://github.com/ollama/ollama/blob/main/docs/gpu.md)
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8316/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8316/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8562
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8562/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8562/comments
|
https://api.github.com/repos/ollama/ollama/issues/8562/events
|
https://github.com/ollama/ollama/issues/8562
| 2,809,089,096
|
I_kwDOJ0Z1Ps6nb0xI
| 8,562
|
None value being returned from with_structured_output request
|
{
"login": "jonmach",
"id": 7843429,
"node_id": "MDQ6VXNlcjc4NDM0Mjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jonmach",
"html_url": "https://github.com/jonmach",
"followers_url": "https://api.github.com/users/jonmach/followers",
"following_url": "https://api.github.com/users/jonmach/following{/other_user}",
"gists_url": "https://api.github.com/users/jonmach/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jonmach/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jonmach/subscriptions",
"organizations_url": "https://api.github.com/users/jonmach/orgs",
"repos_url": "https://api.github.com/users/jonmach/repos",
"events_url": "https://api.github.com/users/jonmach/events{/privacy}",
"received_events_url": "https://api.github.com/users/jonmach/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 6
| 2025-01-24T10:26:05
| 2025-01-27T10:15:42
| 2025-01-24T13:27:20
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Python versions are:
langchain 0.3.15
langchain-community 0.3.15
langchain-core 0.3.31
langchain-ollama 0.2.2
ollama 0.4.7
Running ollama 0.5.7 (pip install -U ollama did not increase the version beyond 0.4.7)
---
Using **with_structured_output()** seems to work for a very simple example such as the following:
```
from langchain_ollama import ChatOllama
from typing import Optional
from pydantic import BaseModel, Field
class Person(BaseModel):
name: str
age: int
llm = ChatOllama(
model="qwen2.5:1.5b",
temperature=0,
).with_structured_output(Person)
llm.invoke("Erick 27")
```
However, for a more complex requirement, it fails with Ollama returning a value of None.
```
from pydantic import BaseModel, Field
from typing import Optional
from openai import OpenAI
from langchain_ollama import ChatOllama
# Define the output model
class Experience(BaseModel):
company: str = Field(..., description="The name of the company.")
position: str = Field(..., description="The job title held at the company.")
start_date: str = Field(..., description="The date when you started working at the company.")
end_date: str = Field(..., description="The date when you left the company. If still employed, use 'Present'.")
class Education(BaseModel):
institution_name: str = Field(..., description="The name of the educational institution.")
degree: str = Field(..., description="The degree obtained from the institution.")
start_date: str = Field(..., description="The date when you started attending school at the institution.")
end_date: str = Field(..., description="The date when you graduated. If still enrolled, use 'Present'.")
class Resume(BaseModel):
full_name: str = Field(..., description="The full name of the person on the resume.")
contact_email: str = Field(..., description="The email address for contacting the person.")
phone_number: str = Field(..., description="The phone number for contacting the person.")
summary: str = Field(..., description="A brief summary of the person's career highlights.")
experience: Optional[list[Experience]] = Field([], description="List of experiences held by the person.")
education: Optional[list[Education]] = Field([], description="List of educational institutions attended by the person.")
with open('CVs/resume.md', 'r') as file:
resume_data = file.read()
verbose=True
model = "qwen2.5:14b"
prompt = f"""
Analyse the following resume from the content between the triple backticks below: For the resume below, identify the following information:
1) Their personal details, including name, email, phone number and anything else they provide.
2) An overall summary of their experience to provide a general background.
3) A list of the companies they have worked for. This should include the company name, the dates they started and and ended working for the company, and the tasks and activities they carried out.
4) A list of universities or colleges that the person went to. This should include the name of the college the title of the qualification, and the dates they started and ended.
The raw data is here:
```{resume_data}```
"""
# print(prompt)
llm = ChatOllama( model=model,
num_ctx = 32000,
timeout = 600,
temperature=0.0,
verboseness = verbose,
response = "json")
structured_llm = llm.with_structured_output(Resume)
print("Calling LLM")
response = structured_llm.invoke(prompt)
print(response)
```
It also fails without the 'response = "json" included.
I just get a None response.
Oddly, this is not consistent. Sometimes, I get back a response, but it fails satisfying the Resume type. because it won't find education items. Even though Education is an optional type in the Resume class.
For some reason, I cannot upload the small resume file, so here it is in cleartext:
---
```
## **Professional Experience**
### **Senior Software Engineer**
**Tech Innovators Inc.**
_June 2015 – Present_
- Designed and implemented scalable microservices architecture for a SaaS platform, improving performance by 30%.
- Led a team of 12 engineers, mentoring junior developers and conducting regular code reviews.
- Integrated AI/ML capabilities into legacy systems, increasing operational efficiency by 20%.
- Championed DevOps practices, reducing deployment times from days to hours.
### **Software Architect**
**NextGen Solutions**
_March 2010 – May 2015_
- Architected and delivered a real-time analytics platform for financial services, handling millions of transactions daily.
- Migrated a monolithic system to a distributed microservices-based architecture, enabling faster feature delivery.
- Partnered with product managers to define technical requirements and roadmap, aligning business goals with engineering efforts.
### **Lead Developer**
**Alpha Development Corp.**
_January 2005 – February 2010_
- Built a high-availability e-commerce platform that handled over 500,000 daily users.
- Created APIs to integrate third-party payment gateways, enhancing user experience and reducing downtime.
- Conducted performance optimizations that improved application speed by 40%.
### **Software Engineer**
**CodeSphere LLC**
_June 2000 – December 2004_
- Developed enterprise-grade web applications using Java and C++.
- Automated internal processes, saving the company 15% in operational costs annually.
- Collaborated with cross-functional teams to deliver projects on time and within budget.
---
## **Education**
### **Master of Science in Computer Science**
**Massachusetts Institute of Technology**
_August 1998 – May 2000_
### **Bachelor of Science in Computer Science**
**University of California, Berkeley**
_August 1994 – May 1998_
---
## **Skills**
- Programming Languages: Python, Java, C++, JavaScript
- Cloud Platforms: AWS, Azure, Google Cloud
- Architecture: Microservices, Distributed Systems, RESTful APIs
- Tools: Docker, Kubernetes, Terraform
- Agile Development, DevOps, AI/ML Integration
---
## **Certifications**
- AWS Certified Solutions Architect – Professional
- Certified Kubernetes Administrator (CKA)
- Certified ScrumMaster (CSM)
---
## **Contact**
Feel free to reach out via email or phone for opportunities or collaboration.
```
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.5.7
|
{
"login": "jonmach",
"id": 7843429,
"node_id": "MDQ6VXNlcjc4NDM0Mjk=",
"avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jonmach",
"html_url": "https://github.com/jonmach",
"followers_url": "https://api.github.com/users/jonmach/followers",
"following_url": "https://api.github.com/users/jonmach/following{/other_user}",
"gists_url": "https://api.github.com/users/jonmach/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jonmach/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jonmach/subscriptions",
"organizations_url": "https://api.github.com/users/jonmach/orgs",
"repos_url": "https://api.github.com/users/jonmach/repos",
"events_url": "https://api.github.com/users/jonmach/events{/privacy}",
"received_events_url": "https://api.github.com/users/jonmach/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8562/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8562/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/130
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/130/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/130/comments
|
https://api.github.com/repos/ollama/ollama/issues/130/events
|
https://github.com/ollama/ollama/pull/130
| 1,812,847,886
|
PR_kwDOJ0Z1Ps5V8NI1
| 130
|
vendor in progress bar and change to bytes instead of bibytes
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-19T22:54:28
| 2023-07-20T00:24:03
| 2023-07-20T00:24:03
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/130",
"html_url": "https://github.com/ollama/ollama/pull/130",
"diff_url": "https://github.com/ollama/ollama/pull/130.diff",
"patch_url": "https://github.com/ollama/ollama/pull/130.patch",
"merged_at": "2023-07-20T00:24:03"
}
| null |
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/130/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/130/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4611
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4611/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4611/comments
|
https://api.github.com/repos/ollama/ollama/issues/4611/events
|
https://github.com/ollama/ollama/issues/4611
| 2,314,861,010
|
I_kwDOJ0Z1Ps6J-fnS
| 4,611
|
User comments on personal model page
|
{
"login": "razvanab",
"id": 2854730,
"node_id": "MDQ6VXNlcjI4NTQ3MzA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2854730?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/razvanab",
"html_url": "https://github.com/razvanab",
"followers_url": "https://api.github.com/users/razvanab/followers",
"following_url": "https://api.github.com/users/razvanab/following{/other_user}",
"gists_url": "https://api.github.com/users/razvanab/gists{/gist_id}",
"starred_url": "https://api.github.com/users/razvanab/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/razvanab/subscriptions",
"organizations_url": "https://api.github.com/users/razvanab/orgs",
"repos_url": "https://api.github.com/users/razvanab/repos",
"events_url": "https://api.github.com/users/razvanab/events{/privacy}",
"received_events_url": "https://api.github.com/users/razvanab/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] |
open
| false
| null |
[] | null | 0
| 2024-05-24T09:05:07
| 2024-07-08T17:22:09
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I would like to suggest a section for user comments on the personal Ollama model page. This will help with better tuning the parameters for your version of the model.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4611/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4611/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/4287
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4287/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4287/comments
|
https://api.github.com/repos/ollama/ollama/issues/4287/events
|
https://github.com/ollama/ollama/pull/4287
| 2,287,796,918
|
PR_kwDOJ0Z1Ps5u_yAg
| 4,287
|
suppress ollama_llama_server.exe blank command window popup
|
{
"login": "ashokgelal",
"id": 401055,
"node_id": "MDQ6VXNlcjQwMTA1NQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/401055?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ashokgelal",
"html_url": "https://github.com/ashokgelal",
"followers_url": "https://api.github.com/users/ashokgelal/followers",
"following_url": "https://api.github.com/users/ashokgelal/following{/other_user}",
"gists_url": "https://api.github.com/users/ashokgelal/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ashokgelal/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ashokgelal/subscriptions",
"organizations_url": "https://api.github.com/users/ashokgelal/orgs",
"repos_url": "https://api.github.com/users/ashokgelal/repos",
"events_url": "https://api.github.com/users/ashokgelal/events{/privacy}",
"received_events_url": "https://api.github.com/users/ashokgelal/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 10
| 2024-05-09T14:19:20
| 2025-01-29T16:31:20
| 2024-11-23T21:29:08
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4287",
"html_url": "https://github.com/ollama/ollama/pull/4287",
"diff_url": "https://github.com/ollama/ollama/pull/4287.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4287.patch",
"merged_at": null
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4287/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4287/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1827
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1827/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1827/comments
|
https://api.github.com/repos/ollama/ollama/issues/1827/events
|
https://github.com/ollama/ollama/issues/1827
| 2,068,731,697
|
I_kwDOJ0Z1Ps57Tlcx
| 1,827
|
Massive slowdown on v 0.1.18 vs 0.1.17 with same model on Intel Mac
|
{
"login": "pjv",
"id": 327716,
"node_id": "MDQ6VXNlcjMyNzcxNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/327716?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pjv",
"html_url": "https://github.com/pjv",
"followers_url": "https://api.github.com/users/pjv/followers",
"following_url": "https://api.github.com/users/pjv/following{/other_user}",
"gists_url": "https://api.github.com/users/pjv/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pjv/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pjv/subscriptions",
"organizations_url": "https://api.github.com/users/pjv/orgs",
"repos_url": "https://api.github.com/users/pjv/repos",
"events_url": "https://api.github.com/users/pjv/events{/privacy}",
"received_events_url": "https://api.github.com/users/pjv/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 8
| 2024-01-06T17:18:27
| 2024-04-15T01:15:54
| 2024-01-07T05:46:19
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
<img width="255" alt="SCR-20240106-kfri" src="https://github.com/jmorganca/ollama/assets/327716/c1b5ce5b-acd5-4c81-a59a-6db39ff6a257">
I don’t have exact timings but the same model (in this case, `deepseek-coder:6.7b-instruct-q4_K_S`) generates tokens roughly 5 times faster on 0.1.17 than on 0.1.18 on my Intel Mac.
I upgraded to 0.1.18 and noticed the slowdown in token generation and then downgraded back to 0.1.17 and immediately saw the faster throughput I am accustomed to.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1827/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1827/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8108
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8108/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8108/comments
|
https://api.github.com/repos/ollama/ollama/issues/8108/events
|
https://github.com/ollama/ollama/issues/8108
| 2,740,794,394
|
I_kwDOJ0Z1Ps6jXTQa
| 8,108
|
Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64
|
{
"login": "dernikolas",
"id": 150927066,
"node_id": "U_kgDOCP722g",
"avatar_url": "https://avatars.githubusercontent.com/u/150927066?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dernikolas",
"html_url": "https://github.com/dernikolas",
"followers_url": "https://api.github.com/users/dernikolas/followers",
"following_url": "https://api.github.com/users/dernikolas/following{/other_user}",
"gists_url": "https://api.github.com/users/dernikolas/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dernikolas/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dernikolas/subscriptions",
"organizations_url": "https://api.github.com/users/dernikolas/orgs",
"repos_url": "https://api.github.com/users/dernikolas/repos",
"events_url": "https://api.github.com/users/dernikolas/events{/privacy}",
"received_events_url": "https://api.github.com/users/dernikolas/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5755339642,
"node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg",
"url": "https://api.github.com/repos/ollama/ollama/labels/linux",
"name": "linux",
"color": "516E70",
"default": false,
"description": ""
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] |
open
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 2
| 2024-12-15T17:57:17
| 2025-01-06T17:54:55
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
When I start llama3.2 with `ollama run llama3.2` this error spawns:
`Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64`
I am on Cachy OS with ollama and ollama-rocm installed from extra repo and a AMD RX6800 is the GPU
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.5.2
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8108/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8108/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/6571
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6571/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6571/comments
|
https://api.github.com/repos/ollama/ollama/issues/6571/events
|
https://github.com/ollama/ollama/issues/6571
| 2,497,977,180
|
I_kwDOJ0Z1Ps6U5Btc
| 6,571
|
Impossible to connect to ollama locally from another pc
|
{
"login": "Wilnox23",
"id": 82610998,
"node_id": "MDQ6VXNlcjgyNjEwOTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Wilnox23",
"html_url": "https://github.com/Wilnox23",
"followers_url": "https://api.github.com/users/Wilnox23/followers",
"following_url": "https://api.github.com/users/Wilnox23/following{/other_user}",
"gists_url": "https://api.github.com/users/Wilnox23/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Wilnox23/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Wilnox23/subscriptions",
"organizations_url": "https://api.github.com/users/Wilnox23/orgs",
"repos_url": "https://api.github.com/users/Wilnox23/repos",
"events_url": "https://api.github.com/users/Wilnox23/events{/privacy}",
"received_events_url": "https://api.github.com/users/Wilnox23/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-08-30T18:45:36
| 2024-08-30T21:12:58
| 2024-08-30T21:11:07
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hello, I'm using ollama on windows but I can't connect to the server from another PC on the local network.
I have already changed the following system environment variables.
`OLLAMA_HOST=0.0.0.0`
`OLLAMA_ORIGINS=*`
Thank you for your assistance.
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.8
|
{
"login": "Wilnox23",
"id": 82610998,
"node_id": "MDQ6VXNlcjgyNjEwOTk4",
"avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Wilnox23",
"html_url": "https://github.com/Wilnox23",
"followers_url": "https://api.github.com/users/Wilnox23/followers",
"following_url": "https://api.github.com/users/Wilnox23/following{/other_user}",
"gists_url": "https://api.github.com/users/Wilnox23/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Wilnox23/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Wilnox23/subscriptions",
"organizations_url": "https://api.github.com/users/Wilnox23/orgs",
"repos_url": "https://api.github.com/users/Wilnox23/repos",
"events_url": "https://api.github.com/users/Wilnox23/events{/privacy}",
"received_events_url": "https://api.github.com/users/Wilnox23/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6571/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6571/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2693
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2693/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2693/comments
|
https://api.github.com/repos/ollama/ollama/issues/2693/events
|
https://github.com/ollama/ollama/issues/2693
| 2,149,835,982
|
I_kwDOJ0Z1Ps6AI-TO
| 2,693
|
Ollama windows version not working
|
{
"login": "sdrshn-nmbr",
"id": 122426494,
"node_id": "U_kgDOB0wUfg",
"avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sdrshn-nmbr",
"html_url": "https://github.com/sdrshn-nmbr",
"followers_url": "https://api.github.com/users/sdrshn-nmbr/followers",
"following_url": "https://api.github.com/users/sdrshn-nmbr/following{/other_user}",
"gists_url": "https://api.github.com/users/sdrshn-nmbr/gists{/gist_id}",
"starred_url": "https://api.github.com/users/sdrshn-nmbr/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sdrshn-nmbr/subscriptions",
"organizations_url": "https://api.github.com/users/sdrshn-nmbr/orgs",
"repos_url": "https://api.github.com/users/sdrshn-nmbr/repos",
"events_url": "https://api.github.com/users/sdrshn-nmbr/events{/privacy}",
"received_events_url": "https://api.github.com/users/sdrshn-nmbr/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-02-22T19:55:55
| 2024-02-22T20:24:32
| 2024-02-22T20:24:12
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
It was working fine even yesterday, but I got an update notification and it hasn't been working since. I even tried deleting and reinstalling the installer exe, but it seems the app shows up for a few seconds and then disappears again, but powershell still recognizes the command - it just says ollama not running.
|
{
"login": "sdrshn-nmbr",
"id": 122426494,
"node_id": "U_kgDOB0wUfg",
"avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sdrshn-nmbr",
"html_url": "https://github.com/sdrshn-nmbr",
"followers_url": "https://api.github.com/users/sdrshn-nmbr/followers",
"following_url": "https://api.github.com/users/sdrshn-nmbr/following{/other_user}",
"gists_url": "https://api.github.com/users/sdrshn-nmbr/gists{/gist_id}",
"starred_url": "https://api.github.com/users/sdrshn-nmbr/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sdrshn-nmbr/subscriptions",
"organizations_url": "https://api.github.com/users/sdrshn-nmbr/orgs",
"repos_url": "https://api.github.com/users/sdrshn-nmbr/repos",
"events_url": "https://api.github.com/users/sdrshn-nmbr/events{/privacy}",
"received_events_url": "https://api.github.com/users/sdrshn-nmbr/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2693/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2693/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5758
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5758/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5758/comments
|
https://api.github.com/repos/ollama/ollama/issues/5758/events
|
https://github.com/ollama/ollama/pull/5758
| 2,414,681,278
|
PR_kwDOJ0Z1Ps51sq18
| 5,758
|
marshal json automatically for some template values
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-07-17T22:02:46
| 2024-07-17T22:35:13
| 2024-07-17T22:35:11
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5758",
"html_url": "https://github.com/ollama/ollama/pull/5758",
"diff_url": "https://github.com/ollama/ollama/pull/5758.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5758.patch",
"merged_at": "2024-07-17T22:35:11"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5758/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5758/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3678
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3678/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3678/comments
|
https://api.github.com/repos/ollama/ollama/issues/3678/events
|
https://github.com/ollama/ollama/pull/3678
| 2,246,645,169
|
PR_kwDOJ0Z1Ps5s1fe1
| 3,678
|
darwin: no partial offloading if required memory greater than system
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-04-16T18:23:02
| 2024-04-16T19:05:57
| 2024-04-16T19:05:57
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3678",
"html_url": "https://github.com/ollama/ollama/pull/3678",
"diff_url": "https://github.com/ollama/ollama/pull/3678.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3678.patch",
"merged_at": "2024-04-16T19:05:56"
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3678/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3678/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4381
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4381/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4381/comments
|
https://api.github.com/repos/ollama/ollama/issues/4381/events
|
https://github.com/ollama/ollama/issues/4381
| 2,291,487,916
|
I_kwDOJ0Z1Ps6IlVSs
| 4,381
|
Yi 1.5 was released
|
{
"login": "ddpasa",
"id": 112642920,
"node_id": "U_kgDOBrbLaA",
"avatar_url": "https://avatars.githubusercontent.com/u/112642920?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ddpasa",
"html_url": "https://github.com/ddpasa",
"followers_url": "https://api.github.com/users/ddpasa/followers",
"following_url": "https://api.github.com/users/ddpasa/following{/other_user}",
"gists_url": "https://api.github.com/users/ddpasa/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ddpasa/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ddpasa/subscriptions",
"organizations_url": "https://api.github.com/users/ddpasa/orgs",
"repos_url": "https://api.github.com/users/ddpasa/repos",
"events_url": "https://api.github.com/users/ddpasa/events{/privacy}",
"received_events_url": "https://api.github.com/users/ddpasa/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-05-12T20:23:39
| 2024-05-14T03:35:51
| 2024-05-13T21:34:04
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
https://github.com/01-ai/Yi-1.5
https://huggingface.co/collections/01-ai/yi-15-2024-05-663f3ecab5f815a3eaca7ca8
|
{
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/users/mchiang0610/followers",
"following_url": "https://api.github.com/users/mchiang0610/following{/other_user}",
"gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions",
"organizations_url": "https://api.github.com/users/mchiang0610/orgs",
"repos_url": "https://api.github.com/users/mchiang0610/repos",
"events_url": "https://api.github.com/users/mchiang0610/events{/privacy}",
"received_events_url": "https://api.github.com/users/mchiang0610/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4381/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4381/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/783
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/783/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/783/comments
|
https://api.github.com/repos/ollama/ollama/issues/783/events
|
https://github.com/ollama/ollama/pull/783
| 1,942,570,643
|
PR_kwDOJ0Z1Ps5cxVPh
| 783
|
fix: offloading on low end GPUs
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-10-13T20:08:46
| 2023-10-13T21:36:45
| 2023-10-13T21:36:44
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/783",
"html_url": "https://github.com/ollama/ollama/pull/783",
"diff_url": "https://github.com/ollama/ollama/pull/783.diff",
"patch_url": "https://github.com/ollama/ollama/pull/783.patch",
"merged_at": "2023-10-13T21:36:44"
}
|
Fixes two issues when using low end GPUs:
GPUs with low VRAM are disproportionately affected by overhead when offloading so any device that has less than 2GB VRAM will be exclusively CPU unless overwritten by num_gpu.
A CUDA-enabled runner will still offload to GPU even if num_gpu is 0. This is problematic when the GPU doesn't support a compatible version of CUDA. In this case, select the CPU runner instead.
Caveat: for MacOS (darwin) `go generate` only builds Metal on ARM so it shouldn't be marked as `Accelerated` since there's no fallback
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/783/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/783/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/861
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/861/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/861/comments
|
https://api.github.com/repos/ollama/ollama/issues/861/events
|
https://github.com/ollama/ollama/pull/861
| 1,954,903,919
|
PR_kwDOJ0Z1Ps5dakP4
| 861
|
Added /clear command to clear the buffer
|
{
"login": "pepperoni21",
"id": 29759371,
"node_id": "MDQ6VXNlcjI5NzU5Mzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pepperoni21",
"html_url": "https://github.com/pepperoni21",
"followers_url": "https://api.github.com/users/pepperoni21/followers",
"following_url": "https://api.github.com/users/pepperoni21/following{/other_user}",
"gists_url": "https://api.github.com/users/pepperoni21/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pepperoni21/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pepperoni21/subscriptions",
"organizations_url": "https://api.github.com/users/pepperoni21/orgs",
"repos_url": "https://api.github.com/users/pepperoni21/repos",
"events_url": "https://api.github.com/users/pepperoni21/events{/privacy}",
"received_events_url": "https://api.github.com/users/pepperoni21/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2023-10-20T19:25:06
| 2023-10-20T20:10:22
| 2023-10-20T20:05:19
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/861",
"html_url": "https://github.com/ollama/ollama/pull/861",
"diff_url": "https://github.com/ollama/ollama/pull/861.diff",
"patch_url": "https://github.com/ollama/ollama/pull/861.patch",
"merged_at": null
}
|
Hey,
I thought a /clear command could enhance user experience.
It should work for both windows and linux/mac.
|
{
"login": "pepperoni21",
"id": 29759371,
"node_id": "MDQ6VXNlcjI5NzU5Mzcx",
"avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pepperoni21",
"html_url": "https://github.com/pepperoni21",
"followers_url": "https://api.github.com/users/pepperoni21/followers",
"following_url": "https://api.github.com/users/pepperoni21/following{/other_user}",
"gists_url": "https://api.github.com/users/pepperoni21/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pepperoni21/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pepperoni21/subscriptions",
"organizations_url": "https://api.github.com/users/pepperoni21/orgs",
"repos_url": "https://api.github.com/users/pepperoni21/repos",
"events_url": "https://api.github.com/users/pepperoni21/events{/privacy}",
"received_events_url": "https://api.github.com/users/pepperoni21/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/861/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/861/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4830
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4830/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4830/comments
|
https://api.github.com/repos/ollama/ollama/issues/4830/events
|
https://github.com/ollama/ollama/issues/4830
| 2,335,295,422
|
I_kwDOJ0Z1Ps6LMce-
| 4,830
|
model import to Ollama
|
{
"login": "fingerfree77",
"id": 124861936,
"node_id": "U_kgDOB3E98A",
"avatar_url": "https://avatars.githubusercontent.com/u/124861936?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fingerfree77",
"html_url": "https://github.com/fingerfree77",
"followers_url": "https://api.github.com/users/fingerfree77/followers",
"following_url": "https://api.github.com/users/fingerfree77/following{/other_user}",
"gists_url": "https://api.github.com/users/fingerfree77/gists{/gist_id}",
"starred_url": "https://api.github.com/users/fingerfree77/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fingerfree77/subscriptions",
"organizations_url": "https://api.github.com/users/fingerfree77/orgs",
"repos_url": "https://api.github.com/users/fingerfree77/repos",
"events_url": "https://api.github.com/users/fingerfree77/events{/privacy}",
"received_events_url": "https://api.github.com/users/fingerfree77/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 3
| 2024-06-05T08:54:55
| 2024-09-04T13:31:21
| 2024-09-04T13:31:21
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Is it possible to import model to Ollama from HuggingFace which is not listed in the script convert-hf-to-gguf.py?
In my case model: https://huggingface.co/ai-forever/ruGPT-3.5-13B
When I try import got error:
```
Traceback (most recent call last):
File "llm/llama.cpp/convert-hf-to-gguf.py", line 2865, in <module>
main()
File "llm/llama.cpp/convert-hf-to-gguf.py", line 2850, in main
model_instance.set_vocab()
File "llm/llama.cpp/convert-hf-to-gguf.py", line 114, in set_vocab
self._set_vocab_gpt2()
File "llm/llama.cpp/convert-hf-to-gguf.py", line 500, in _set_vocab_gpt2
tokens, toktypes, tokpre = self.get_vocab_base()
File "llm/llama.cpp/convert-hf-to-gguf.py", line 379, in get_vocab_base
tokpre = self.get_vocab_base_pre(tokenizer)
File "llm/llama.cpp/convert-hf-to-gguf.py", line 491, in get_vocab_base_pre
raise NotImplementedError("BPE pre-tokenizer was not recognized - update get_vocab_base_pre()")
NotImplementedError: BPE pre-tokenizer was not recognized - update get_vocab_base_pre()
```
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4830/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4830/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3569
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3569/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3569/comments
|
https://api.github.com/repos/ollama/ollama/issues/3569/events
|
https://github.com/ollama/ollama/issues/3569
| 2,234,581,592
|
I_kwDOJ0Z1Ps6FMQJY
| 3,569
|
Could anyone tell me the purpose of "ollama-linux-amd64-rocm.tgz" in the released files?
|
{
"login": "airforce011",
"id": 36444699,
"node_id": "MDQ6VXNlcjM2NDQ0Njk5",
"avatar_url": "https://avatars.githubusercontent.com/u/36444699?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/airforce011",
"html_url": "https://github.com/airforce011",
"followers_url": "https://api.github.com/users/airforce011/followers",
"following_url": "https://api.github.com/users/airforce011/following{/other_user}",
"gists_url": "https://api.github.com/users/airforce011/gists{/gist_id}",
"starred_url": "https://api.github.com/users/airforce011/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/airforce011/subscriptions",
"organizations_url": "https://api.github.com/users/airforce011/orgs",
"repos_url": "https://api.github.com/users/airforce011/repos",
"events_url": "https://api.github.com/users/airforce011/events{/privacy}",
"received_events_url": "https://api.github.com/users/airforce011/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1
| 2024-04-10T01:29:39
| 2024-04-23T00:02:08
| 2024-04-23T00:02:08
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What are you trying to do?
I'm trying to run Ollama on Linux (Ubuntu 22.04) with ROCm, but I'm unable to use it even after following the installation instructions in the documentation. Do I need to use the "ollama-linux-amd64-rocm.tgz" file? However, I'm not sure how to use this file, and there are no instructions provided. Any help would be greatly appreciated!
### How should we solve this?
_No response_
### What is the impact of not solving this?
_No response_
### Anything else?
_No response_
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3569/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3569/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4963
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4963/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4963/comments
|
https://api.github.com/repos/ollama/ollama/issues/4963/events
|
https://github.com/ollama/ollama/issues/4963
| 2,344,382,940
|
I_kwDOJ0Z1Ps6LvHHc
| 4,963
|
Ollama in private EC2
|
{
"login": "arpitdish",
"id": 123781157,
"node_id": "U_kgDOB2DAJQ",
"avatar_url": "https://avatars.githubusercontent.com/u/123781157?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/arpitdish",
"html_url": "https://github.com/arpitdish",
"followers_url": "https://api.github.com/users/arpitdish/followers",
"following_url": "https://api.github.com/users/arpitdish/following{/other_user}",
"gists_url": "https://api.github.com/users/arpitdish/gists{/gist_id}",
"starred_url": "https://api.github.com/users/arpitdish/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/arpitdish/subscriptions",
"organizations_url": "https://api.github.com/users/arpitdish/orgs",
"repos_url": "https://api.github.com/users/arpitdish/repos",
"events_url": "https://api.github.com/users/arpitdish/events{/privacy}",
"received_events_url": "https://api.github.com/users/arpitdish/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-06-10T16:40:51
| 2024-07-10T23:01:21
| 2024-07-10T23:01:21
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
How to run ollama on private EC2
### OS
Linux
### GPU
Nvidia
### CPU
_No response_
### Ollama version
_No response_
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4963/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4963/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4290
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4290/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4290/comments
|
https://api.github.com/repos/ollama/ollama/issues/4290/events
|
https://github.com/ollama/ollama/pull/4290
| 2,288,039,651
|
PR_kwDOJ0Z1Ps5vAm1e
| 4,290
|
fix(server): prune files
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-05-09T16:25:11
| 2024-11-23T21:28:14
| 2024-11-23T21:28:14
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4290",
"html_url": "https://github.com/ollama/ollama/pull/4290",
"diff_url": "https://github.com/ollama/ollama/pull/4290.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4290.patch",
"merged_at": null
}
|
follow up to #4269
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4290/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4290/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5627
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5627/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5627/comments
|
https://api.github.com/repos/ollama/ollama/issues/5627/events
|
https://github.com/ollama/ollama/issues/5627
| 2,402,504,053
|
I_kwDOJ0Z1Ps6PM011
| 5,627
|
unknown model architecture: 'gemma2'
|
{
"login": "mouse33333",
"id": 4304174,
"node_id": "MDQ6VXNlcjQzMDQxNzQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/4304174?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mouse33333",
"html_url": "https://github.com/mouse33333",
"followers_url": "https://api.github.com/users/mouse33333/followers",
"following_url": "https://api.github.com/users/mouse33333/following{/other_user}",
"gists_url": "https://api.github.com/users/mouse33333/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mouse33333/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mouse33333/subscriptions",
"organizations_url": "https://api.github.com/users/mouse33333/orgs",
"repos_url": "https://api.github.com/users/mouse33333/repos",
"events_url": "https://api.github.com/users/mouse33333/events{/privacy}",
"received_events_url": "https://api.github.com/users/mouse33333/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-07-11T07:47:19
| 2024-07-11T08:02:42
| 2024-07-11T07:55:05
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I'm trying to run models on my internal Intel GPU.
Here's the software env:
1. activate a virtual env in Anaconda
2. already installed ipex-llm[cpp], ipex-llm[xpu]
3. start the Ollama serve
4. start gemma2 model
Then the log said: unknown model architecture: 'gemma2'
Here's the sreenshot of the log:

### OS
Windows, Docker
### GPU
Intel
### CPU
Intel
### Ollama version
0.2.1
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5627/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5627/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4011
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4011/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4011/comments
|
https://api.github.com/repos/ollama/ollama/issues/4011/events
|
https://github.com/ollama/ollama/issues/4011
| 2,267,922,429
|
I_kwDOJ0Z1Ps6HLb_9
| 4,011
|
When my ollama has both the gemma and llama large models, how can I enable them at the same time?Thank you.
|
{
"login": "joylijoy",
"id": 167687386,
"node_id": "U_kgDOCf602g",
"avatar_url": "https://avatars.githubusercontent.com/u/167687386?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joylijoy",
"html_url": "https://github.com/joylijoy",
"followers_url": "https://api.github.com/users/joylijoy/followers",
"following_url": "https://api.github.com/users/joylijoy/following{/other_user}",
"gists_url": "https://api.github.com/users/joylijoy/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joylijoy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joylijoy/subscriptions",
"organizations_url": "https://api.github.com/users/joylijoy/orgs",
"repos_url": "https://api.github.com/users/joylijoy/repos",
"events_url": "https://api.github.com/users/joylijoy/events{/privacy}",
"received_events_url": "https://api.github.com/users/joylijoy/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-04-29T00:54:44
| 2024-05-01T04:56:32
| 2024-05-01T04:56:32
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
When my ollama has both the gemma and llama large models, how can I enable them at the same time? Does pressing CTRL+D mean to exit the large model? Thank you.
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4011/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4011/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8457
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8457/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8457/comments
|
https://api.github.com/repos/ollama/ollama/issues/8457/events
|
https://github.com/ollama/ollama/issues/8457
| 2,792,336,253
|
I_kwDOJ0Z1Ps6mb6t9
| 8,457
|
model wanted in ollama please:Qwen2.5-Math-PRM-7B
|
{
"login": "twythebest",
"id": 89891289,
"node_id": "MDQ6VXNlcjg5ODkxMjg5",
"avatar_url": "https://avatars.githubusercontent.com/u/89891289?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/twythebest",
"html_url": "https://github.com/twythebest",
"followers_url": "https://api.github.com/users/twythebest/followers",
"following_url": "https://api.github.com/users/twythebest/following{/other_user}",
"gists_url": "https://api.github.com/users/twythebest/gists{/gist_id}",
"starred_url": "https://api.github.com/users/twythebest/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/twythebest/subscriptions",
"organizations_url": "https://api.github.com/users/twythebest/orgs",
"repos_url": "https://api.github.com/users/twythebest/repos",
"events_url": "https://api.github.com/users/twythebest/events{/privacy}",
"received_events_url": "https://api.github.com/users/twythebest/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
open
| false
| null |
[] | null | 0
| 2025-01-16T10:36:00
| 2025-01-16T10:36:00
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
model wanted in ollama please:Qwen2.5-Math-PRM-7B
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8457/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8457/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/6544
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6544/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6544/comments
|
https://api.github.com/repos/ollama/ollama/issues/6544/events
|
https://github.com/ollama/ollama/issues/6544
| 2,492,824,185
|
I_kwDOJ0Z1Ps6UlXp5
| 6,544
|
Specifying options via openai client extra_body are not handled by ollama
|
{
"login": "gaardhus",
"id": 46934916,
"node_id": "MDQ6VXNlcjQ2OTM0OTE2",
"avatar_url": "https://avatars.githubusercontent.com/u/46934916?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gaardhus",
"html_url": "https://github.com/gaardhus",
"followers_url": "https://api.github.com/users/gaardhus/followers",
"following_url": "https://api.github.com/users/gaardhus/following{/other_user}",
"gists_url": "https://api.github.com/users/gaardhus/gists{/gist_id}",
"starred_url": "https://api.github.com/users/gaardhus/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gaardhus/subscriptions",
"organizations_url": "https://api.github.com/users/gaardhus/orgs",
"repos_url": "https://api.github.com/users/gaardhus/repos",
"events_url": "https://api.github.com/users/gaardhus/events{/privacy}",
"received_events_url": "https://api.github.com/users/gaardhus/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 5
| 2024-08-28T19:19:11
| 2024-09-05T21:47:15
| 2024-09-04T13:25:17
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
So I've been trying to set the num_ctx for the mistral-nemo through the openai api client, however it does not seem to have an effect.
```python
# Ollama client
client = OllamaAsyncClient(host=base_url, **kwargs)
chat_completion = await client.chat(
messages=messages,
model=model,
stream=stream,
options={
"num_ctx": 64_000,
"temperature": temperature,
"stop": stop_tokens,
},
**kwargs,
)
message = chat_completion["message"]["content"].strip()
# OpenAI client
client = AsyncOpenAI(api_key=api_key, base_url=base_url, **kwargs)
chat_completion = await client.chat.completions.create(
messages=messages,
model=model,
temperature=temperature,
stream=stream,
stop=stop_tokens,
extra_body={"options": {"num_ctx": 64_000}},
**kwargs,
)
message = chat_completion.choices[0].message.content.strip()
```
Doing it with ollama works, but the extra_body argument from openai seems to be handled as an extra field rather than merged with the rest of the request.
```
'model': 'mistral-nemo:12b-instruct-2407-q8_0', 'stop': None, 'stream': False, 'temperature': 0.7}, 'extra_json': {'options': {'num_ctx': 64000}}}
```
I guess the solution would be to unpack the extra_json field on the server end?
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.8
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6544/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6544/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7106
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7106/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7106/comments
|
https://api.github.com/repos/ollama/ollama/issues/7106/events
|
https://github.com/ollama/ollama/issues/7106
| 2,568,514,399
|
I_kwDOJ0Z1Ps6ZGGtf
| 7,106
|
Update /api/show endpoint to return a boolean describing if the model supports FIM
|
{
"login": "tomasz-stefaniak",
"id": 7911227,
"node_id": "MDQ6VXNlcjc5MTEyMjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/7911227?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tomasz-stefaniak",
"html_url": "https://github.com/tomasz-stefaniak",
"followers_url": "https://api.github.com/users/tomasz-stefaniak/followers",
"following_url": "https://api.github.com/users/tomasz-stefaniak/following{/other_user}",
"gists_url": "https://api.github.com/users/tomasz-stefaniak/gists{/gist_id}",
"starred_url": "https://api.github.com/users/tomasz-stefaniak/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/tomasz-stefaniak/subscriptions",
"organizations_url": "https://api.github.com/users/tomasz-stefaniak/orgs",
"repos_url": "https://api.github.com/users/tomasz-stefaniak/repos",
"events_url": "https://api.github.com/users/tomasz-stefaniak/events{/privacy}",
"received_events_url": "https://api.github.com/users/tomasz-stefaniak/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-10-06T07:32:55
| 2024-11-17T14:49:48
| 2024-11-17T14:49:48
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi,
We support Ollama as a [model provider in Continue.dev](https://docs.continue.dev/customize/model-providers/ollama) and have recently added FIM (Fill-in-the-Middle) support. Currently, there is no API we are aware of that can be used to determine if a model supports FIM, so we guess its capabilities by checking for a ".Suffix" variable in the template returned from Ollama's `/api/show` endpoint.
To improve accuracy and stability, we would appreciate receiving this information directly as a boolean value from the `/api/show` endpoint. This would allow us to integrate FIM support more reliably and provide better service to Continue users.
Context: [GitHub Link](https://github.com/continuedev/continue/pull/2452/files#diff-3f4bdaca0060eb741057719bcc865486ab770e4ea1659b317f77235adc887c37R81)
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7106/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7106/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/6749
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6749/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6749/comments
|
https://api.github.com/repos/ollama/ollama/issues/6749/events
|
https://github.com/ollama/ollama/issues/6749
| 2,519,216,246
|
I_kwDOJ0Z1Ps6WKDB2
| 6,749
|
Add version when the docker container is starting
|
{
"login": "svaningelgem",
"id": 199434,
"node_id": "MDQ6VXNlcjE5OTQzNA==",
"avatar_url": "https://avatars.githubusercontent.com/u/199434?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/svaningelgem",
"html_url": "https://github.com/svaningelgem",
"followers_url": "https://api.github.com/users/svaningelgem/followers",
"following_url": "https://api.github.com/users/svaningelgem/following{/other_user}",
"gists_url": "https://api.github.com/users/svaningelgem/gists{/gist_id}",
"starred_url": "https://api.github.com/users/svaningelgem/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/svaningelgem/subscriptions",
"organizations_url": "https://api.github.com/users/svaningelgem/orgs",
"repos_url": "https://api.github.com/users/svaningelgem/repos",
"events_url": "https://api.github.com/users/svaningelgem/events{/privacy}",
"received_events_url": "https://api.github.com/users/svaningelgem/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-09-11T09:46:48
| 2024-09-12T03:50:22
| 2024-09-11T23:56:59
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
It'd be nice to know what version is running when sharing logs. So my feature request is to add the current version when you start the docker container.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6749/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6749/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4271
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4271/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4271/comments
|
https://api.github.com/repos/ollama/ollama/issues/4271/events
|
https://github.com/ollama/ollama/issues/4271
| 2,286,720,880
|
I_kwDOJ0Z1Ps6ITJdw
| 4,271
|
Partial pruning does not wrok
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 0
| 2024-05-09T01:17:56
| 2024-05-09T01:17:56
| null |
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
```
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:cfcf93119280c4a10c1df57335bad341e000cabbc4faff125531d941a5b0befa:partial:13': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:cfcf93119280c4a10c1df57335bad341e000cabbc4faff125531d941a5b0befa:partial:67': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:4fe022a8902336d3c452c88f7aca5590f5b5b02ccfd06320fdefab02412e1f0b:partial:18': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:431800fe7a30f378018e76cdec8ed5f99e6f81b22fc5f63b0486e326f0b25545:partial:37': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:4fe022a8902336d3c452c88f7aca5590f5b5b02ccfd06320fdefab02412e1f0b:partial:23': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:4fe022a8902336d3c452c88f7aca5590f5b5b02ccfd06320fdefab02412e1f0b:partial:8': invalid digest format"
time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:704 msg="total unused blobs removed: 190"
```
### OS
_No response_
### GPU
_No response_
### CPU
_No response_
### Ollama version
_No response_
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4271/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4271/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/8649
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8649/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8649/comments
|
https://api.github.com/repos/ollama/ollama/issues/8649/events
|
https://github.com/ollama/ollama/issues/8649
| 2,817,207,537
|
I_kwDOJ0Z1Ps6n6yzx
| 8,649
|
Short run response duration calculations are off
|
{
"login": "NerdyShawn",
"id": 16088118,
"node_id": "MDQ6VXNlcjE2MDg4MTE4",
"avatar_url": "https://avatars.githubusercontent.com/u/16088118?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/NerdyShawn",
"html_url": "https://github.com/NerdyShawn",
"followers_url": "https://api.github.com/users/NerdyShawn/followers",
"following_url": "https://api.github.com/users/NerdyShawn/following{/other_user}",
"gists_url": "https://api.github.com/users/NerdyShawn/gists{/gist_id}",
"starred_url": "https://api.github.com/users/NerdyShawn/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/NerdyShawn/subscriptions",
"organizations_url": "https://api.github.com/users/NerdyShawn/orgs",
"repos_url": "https://api.github.com/users/NerdyShawn/repos",
"events_url": "https://api.github.com/users/NerdyShawn/events{/privacy}",
"received_events_url": "https://api.github.com/users/NerdyShawn/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 3
| 2025-01-29T04:22:55
| 2025-01-29T14:03:03
| 2025-01-29T14:03:01
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Running the smaller `deepseek-r1:1.5b` model it seems like the very short time and duration how it is being calculated is off given the duration in the response. It seems since its close to zero it throws off how the time is measured.
---

---
```json
date && time curl -s https://ollama.somecooldomain.lan/api/generate -d '{
"model": "deepseek-r1:1.5b",
"prompt": "What is the meaning of life?",
"stream": false
}' | jq
Tue Jan 28 11:10:42 PM EST 2025
{
"model": "deepseek-r1:1.5b",
"created_at": "2025-01-29T04:10:42.569719236Z",
"response": "<think>\n\n</think>\n\nI am sorry, I cannot answer that question.",
"done": true,
"done_reason": "stop",
"context": [
151644,
3838,
374,
279,
7290,
315,
2272,
30,
151645,
151648,
271,
151649,
271,
40,
1079,
14589,
11,
358,
4157,
4226,
429,
3405,
13
],
"total_duration": 256103089,
"load_duration": 144586102,
"prompt_eval_count": 10,
"prompt_eval_duration": 24000000,
"eval_count": 15,
"eval_duration": 86000000
}
real 0m0.345s
user 0m0.050s
sys 0m0.016s
```
The real time it took to execute was under a half second but the durations are all wrong.
### OS
Docker
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.5.6-0-g2539f2d-dirty
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8649/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8649/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2864
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2864/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2864/comments
|
https://api.github.com/repos/ollama/ollama/issues/2864/events
|
https://github.com/ollama/ollama/issues/2864
| 2,163,502,283
|
I_kwDOJ0Z1Ps6A9GzL
| 2,864
|
Read and write files (To allow for more types of AI models in Ollama)
|
{
"login": "trymeouteh",
"id": 31172274,
"node_id": "MDQ6VXNlcjMxMTcyMjc0",
"avatar_url": "https://avatars.githubusercontent.com/u/31172274?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trymeouteh",
"html_url": "https://github.com/trymeouteh",
"followers_url": "https://api.github.com/users/trymeouteh/followers",
"following_url": "https://api.github.com/users/trymeouteh/following{/other_user}",
"gists_url": "https://api.github.com/users/trymeouteh/gists{/gist_id}",
"starred_url": "https://api.github.com/users/trymeouteh/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/trymeouteh/subscriptions",
"organizations_url": "https://api.github.com/users/trymeouteh/orgs",
"repos_url": "https://api.github.com/users/trymeouteh/repos",
"events_url": "https://api.github.com/users/trymeouteh/events{/privacy}",
"received_events_url": "https://api.github.com/users/trymeouteh/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 4
| 2024-03-01T14:12:41
| 2024-03-08T14:09:08
| 2024-03-07T18:54:15
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
To my understanding it will seem Ollama can only support text chat based models and to my understand it could support more kinds of models by adding a reading files and writing files feature...
- A way to have a model read one or more file on your system and be able to send it to the AI model to process.
- A way for the AI model to return a file, by generating a file (Writing a file) onto the system and them being able to link it into the chat response.
To my understanding, this can allow Ollama to work with more model types such as text to image models, image to image models, text to audio models, audio to text models, audio to audio models, etc.
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2864/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2864/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/502
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/502/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/502/comments
|
https://api.github.com/repos/ollama/ollama/issues/502/events
|
https://github.com/ollama/ollama/issues/502
| 1,889,051,436
|
I_kwDOJ0Z1Ps5wmKMs
| 502
|
Better Document /api/embeddings
|
{
"login": "FairyTail2000",
"id": 22645621,
"node_id": "MDQ6VXNlcjIyNjQ1NjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FairyTail2000",
"html_url": "https://github.com/FairyTail2000",
"followers_url": "https://api.github.com/users/FairyTail2000/followers",
"following_url": "https://api.github.com/users/FairyTail2000/following{/other_user}",
"gists_url": "https://api.github.com/users/FairyTail2000/gists{/gist_id}",
"starred_url": "https://api.github.com/users/FairyTail2000/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/FairyTail2000/subscriptions",
"organizations_url": "https://api.github.com/users/FairyTail2000/orgs",
"repos_url": "https://api.github.com/users/FairyTail2000/repos",
"events_url": "https://api.github.com/users/FairyTail2000/events{/privacy}",
"received_events_url": "https://api.github.com/users/FairyTail2000/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 5
| 2023-09-10T08:37:43
| 2023-10-26T00:12:03
| 2023-10-26T00:12:03
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
It's currently not very clear on how to use is and how it's relation is to the EMBED Modelfile instruction.
Can you enlighten me and/or update the wiki and examples?
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/502/reactions",
"total_count": 4,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 2
}
|
https://api.github.com/repos/ollama/ollama/issues/502/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7634
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7634/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7634/comments
|
https://api.github.com/repos/ollama/ollama/issues/7634/events
|
https://github.com/ollama/ollama/issues/7634
| 2,653,045,312
|
I_kwDOJ0Z1Ps6eIkJA
| 7,634
|
Custom tool are not getting called in tool call
|
{
"login": "Vikneshkumarmohan",
"id": 71374469,
"node_id": "MDQ6VXNlcjcxMzc0NDY5",
"avatar_url": "https://avatars.githubusercontent.com/u/71374469?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Vikneshkumarmohan",
"html_url": "https://github.com/Vikneshkumarmohan",
"followers_url": "https://api.github.com/users/Vikneshkumarmohan/followers",
"following_url": "https://api.github.com/users/Vikneshkumarmohan/following{/other_user}",
"gists_url": "https://api.github.com/users/Vikneshkumarmohan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Vikneshkumarmohan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Vikneshkumarmohan/subscriptions",
"organizations_url": "https://api.github.com/users/Vikneshkumarmohan/orgs",
"repos_url": "https://api.github.com/users/Vikneshkumarmohan/repos",
"events_url": "https://api.github.com/users/Vikneshkumarmohan/events{/privacy}",
"received_events_url": "https://api.github.com/users/Vikneshkumarmohan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-11-12T18:51:44
| 2024-12-03T03:55:02
| 2024-12-02T08:01:16
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
when using ChatOllama in Langgraph tool call is not happening for the custom tools , but it is working as expected with tools like tavily search ,
https://langchain-ai.github.io/langgraph/tutorials/multi_agent/multi-agent-collaboration/#invoke
above is the example for reference code , when working with ollama LLM the custom tool is not calling (custom tool : python_repl),
### OS
Windows
### GPU
AMD
### CPU
AMD
### Ollama version
Llama3.2
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7634/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
}
|
https://api.github.com/repos/ollama/ollama/issues/7634/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6186
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6186/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6186/comments
|
https://api.github.com/repos/ollama/ollama/issues/6186/events
|
https://github.com/ollama/ollama/pull/6186
| 2,449,377,760
|
PR_kwDOJ0Z1Ps53fc2u
| 6,186
|
Implement linux NUMA detection
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-08-05T20:01:08
| 2024-08-05T22:20:11
| 2024-08-05T22:20:07
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/6186",
"html_url": "https://github.com/ollama/ollama/pull/6186",
"diff_url": "https://github.com/ollama/ollama/pull/6186.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6186.patch",
"merged_at": "2024-08-05T22:20:07"
}
|
If the system has multiple numa nodes, enable numa support in llama.cpp If we detect `numactl` in the path, use that, else use the basic "distribute" mode.
This also removes the `use_numa` setting as the bool type is no longer useful given this parameter now requires a mode.
Fixes #6093
Fixes #2496
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6186/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6186/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5715
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5715/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5715/comments
|
https://api.github.com/repos/ollama/ollama/issues/5715/events
|
https://github.com/ollama/ollama/issues/5715
| 2,410,125,004
|
I_kwDOJ0Z1Ps6Pp5bM
| 5,715
|
Allow using `"""` in TEMPLATE Modelfile command
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-07-16T03:33:31
| 2024-07-16T03:33:31
| null |
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null | null | null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5715/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5715/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/7121
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7121/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7121/comments
|
https://api.github.com/repos/ollama/ollama/issues/7121/events
|
https://github.com/ollama/ollama/pull/7121
| 2,571,261,705
|
PR_kwDOJ0Z1Ps5925z4
| 7,121
|
llama: refine developer docs
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-10-07T19:18:10
| 2024-10-07T19:43:49
| 2024-10-07T19:43:46
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7121",
"html_url": "https://github.com/ollama/ollama/pull/7121",
"diff_url": "https://github.com/ollama/ollama/pull/7121.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7121.patch",
"merged_at": "2024-10-07T19:43:46"
}
| null |
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7121/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7121/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8183
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8183/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8183/comments
|
https://api.github.com/repos/ollama/ollama/issues/8183/events
|
https://github.com/ollama/ollama/issues/8183
| 2,752,749,059
|
I_kwDOJ0Z1Ps6kE54D
| 8,183
|
How do I specify specific GPUs when running a model?
|
{
"login": "any35",
"id": 2082210,
"node_id": "MDQ6VXNlcjIwODIyMTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2082210?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/any35",
"html_url": "https://github.com/any35",
"followers_url": "https://api.github.com/users/any35/followers",
"following_url": "https://api.github.com/users/any35/following{/other_user}",
"gists_url": "https://api.github.com/users/any35/gists{/gist_id}",
"starred_url": "https://api.github.com/users/any35/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/any35/subscriptions",
"organizations_url": "https://api.github.com/users/any35/orgs",
"repos_url": "https://api.github.com/users/any35/repos",
"events_url": "https://api.github.com/users/any35/events{/privacy}",
"received_events_url": "https://api.github.com/users/any35/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-12-20T13:30:03
| 2025-01-13T01:43:05
| 2025-01-13T01:43:04
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I have a server with 16 RTX4090 graphics cards installed.
Now I need to run different models at the same time and specify the corresponding GPUs. How can I do this?
For example: run llama 3.3 on GPU0,1; llama 3.2 on gpu 2;...
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8183/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8183/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1306
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1306/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1306/comments
|
https://api.github.com/repos/ollama/ollama/issues/1306/events
|
https://github.com/ollama/ollama/pull/1306
| 2,014,904,973
|
PR_kwDOJ0Z1Ps5glFGZ
| 1,306
|
Fix Docker image not using GPU
|
{
"login": "bwest2397",
"id": 73724178,
"node_id": "MDQ6VXNlcjczNzI0MTc4",
"avatar_url": "https://avatars.githubusercontent.com/u/73724178?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bwest2397",
"html_url": "https://github.com/bwest2397",
"followers_url": "https://api.github.com/users/bwest2397/followers",
"following_url": "https://api.github.com/users/bwest2397/following{/other_user}",
"gists_url": "https://api.github.com/users/bwest2397/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bwest2397/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bwest2397/subscriptions",
"organizations_url": "https://api.github.com/users/bwest2397/orgs",
"repos_url": "https://api.github.com/users/bwest2397/repos",
"events_url": "https://api.github.com/users/bwest2397/events{/privacy}",
"received_events_url": "https://api.github.com/users/bwest2397/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 11
| 2023-11-28T17:06:07
| 2023-12-12T19:34:56
| 2023-11-28T21:25:23
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1306",
"html_url": "https://github.com/ollama/ollama/pull/1306",
"diff_url": "https://github.com/ollama/ollama/pull/1306.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1306.patch",
"merged_at": null
}
|
As it currently stands, the Docker image that gets built is seemingly unable to use the GPU despite the initial "build" stage of the Dockerfile ostensibly being built with CUDA support (i.e. built using the `nvidia/cuda` base image). As reported in https://github.com/jmorganca/ollama/issues/797, it seems that simply setting the second stage of the Dockerfile to use this same `nvidia/cuda` base image resolves the problem.
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1306/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1306/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5667
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5667/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5667/comments
|
https://api.github.com/repos/ollama/ollama/issues/5667/events
|
https://github.com/ollama/ollama/pull/5667
| 2,406,750,057
|
PR_kwDOJ0Z1Ps51SJUt
| 5,667
|
Enhance list command
|
{
"login": "kaichen",
"id": 4048,
"node_id": "MDQ6VXNlcjQwNDg=",
"avatar_url": "https://avatars.githubusercontent.com/u/4048?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kaichen",
"html_url": "https://github.com/kaichen",
"followers_url": "https://api.github.com/users/kaichen/followers",
"following_url": "https://api.github.com/users/kaichen/following{/other_user}",
"gists_url": "https://api.github.com/users/kaichen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kaichen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kaichen/subscriptions",
"organizations_url": "https://api.github.com/users/kaichen/orgs",
"repos_url": "https://api.github.com/users/kaichen/repos",
"events_url": "https://api.github.com/users/kaichen/events{/privacy}",
"received_events_url": "https://api.github.com/users/kaichen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-07-13T07:21:13
| 2024-11-22T00:49:46
| 2024-11-22T00:49:46
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5667",
"html_url": "https://github.com/ollama/ollama/pull/5667",
"diff_url": "https://github.com/ollama/ollama/pull/5667.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5667.patch",
"merged_at": null
}
|
changes
- add missing filter description for list command
- add sort option to list command
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5667/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5667/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/24
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/24/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/24/comments
|
https://api.github.com/repos/ollama/ollama/issues/24/events
|
https://github.com/ollama/ollama/pull/24
| 1,781,626,501
|
PR_kwDOJ0Z1Ps5USMRS
| 24
|
remove add cmd
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-06-29T23:20:02
| 2023-06-29T23:29:30
| 2023-06-29T23:29:23
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/24",
"html_url": "https://github.com/ollama/ollama/pull/24",
"diff_url": "https://github.com/ollama/ollama/pull/24.diff",
"patch_url": "https://github.com/ollama/ollama/pull/24.patch",
"merged_at": "2023-06-29T23:29:23"
}
|
🧹
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/24/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/24/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2417
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2417/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2417/comments
|
https://api.github.com/repos/ollama/ollama/issues/2417/events
|
https://github.com/ollama/ollama/issues/2417
| 2,126,159,957
|
I_kwDOJ0Z1Ps5-uqBV
| 2,417
|
Tools support in ChatCompletion endpoints
|
{
"login": "pamelafox",
"id": 297042,
"node_id": "MDQ6VXNlcjI5NzA0Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/297042?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pamelafox",
"html_url": "https://github.com/pamelafox",
"followers_url": "https://api.github.com/users/pamelafox/followers",
"following_url": "https://api.github.com/users/pamelafox/following{/other_user}",
"gists_url": "https://api.github.com/users/pamelafox/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pamelafox/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pamelafox/subscriptions",
"organizations_url": "https://api.github.com/users/pamelafox/orgs",
"repos_url": "https://api.github.com/users/pamelafox/repos",
"events_url": "https://api.github.com/users/pamelafox/events{/privacy}",
"received_events_url": "https://api.github.com/users/pamelafox/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6657611864,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjNMYWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/compatibility",
"name": "compatibility",
"color": "bfdadc",
"default": false,
"description": ""
}
] |
closed
| false
|
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers",
"following_url": "https://api.github.com/users/bmizerany/following{/other_user}",
"gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions",
"organizations_url": "https://api.github.com/users/bmizerany/orgs",
"repos_url": "https://api.github.com/users/bmizerany/repos",
"events_url": "https://api.github.com/users/bmizerany/events{/privacy}",
"received_events_url": "https://api.github.com/users/bmizerany/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers",
"following_url": "https://api.github.com/users/bmizerany/following{/other_user}",
"gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions",
"organizations_url": "https://api.github.com/users/bmizerany/orgs",
"repos_url": "https://api.github.com/users/bmizerany/repos",
"events_url": "https://api.github.com/users/bmizerany/events{/privacy}",
"received_events_url": "https://api.github.com/users/bmizerany/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 10
| 2024-02-08T22:17:27
| 2024-06-25T04:30:24
| 2024-06-25T04:30:24
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
We'd love tools support so we can use ollama with our existing OpenAI-using apps. Not sure if that's possible across the board with all models.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2417/reactions",
"total_count": 40,
"+1": 40,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2417/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3265
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3265/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3265/comments
|
https://api.github.com/repos/ollama/ollama/issues/3265/events
|
https://github.com/ollama/ollama/issues/3265
| 2,197,028,474
|
I_kwDOJ0Z1Ps6C8_56
| 3,265
|
Does ollama also plan to support the sound models?
|
{
"login": "insooneelife",
"id": 8437769,
"node_id": "MDQ6VXNlcjg0Mzc3Njk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/insooneelife",
"html_url": "https://github.com/insooneelife",
"followers_url": "https://api.github.com/users/insooneelife/followers",
"following_url": "https://api.github.com/users/insooneelife/following{/other_user}",
"gists_url": "https://api.github.com/users/insooneelife/gists{/gist_id}",
"starred_url": "https://api.github.com/users/insooneelife/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/insooneelife/subscriptions",
"organizations_url": "https://api.github.com/users/insooneelife/orgs",
"repos_url": "https://api.github.com/users/insooneelife/repos",
"events_url": "https://api.github.com/users/insooneelife/events{/privacy}",
"received_events_url": "https://api.github.com/users/insooneelife/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 2
| 2024-03-20T09:18:59
| 2024-12-25T15:57:35
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What are you trying to do?
Currently, ollama supports most llms and I know that it also supports vision model. I thought it would be nice if the sound model could also be used through ollama. I wonder if there are any plans like this.
### How should we solve this?
I know that there are currently sound models released on huggingface. There are several TTS and STT models released as open source. I think these models could also support model serving and API like llm.
### What is the impact of not solving this?
_No response_
### Anything else?
_No response_
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3265/reactions",
"total_count": 7,
"+1": 7,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3265/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/6098
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6098/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6098/comments
|
https://api.github.com/repos/ollama/ollama/issues/6098/events
|
https://github.com/ollama/ollama/issues/6098
| 2,440,088,309
|
I_kwDOJ0Z1Ps6RcMr1
| 6,098
|
Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest.
|
{
"login": "fanjikang",
"id": 146172937,
"node_id": "U_kgDOCLZsCQ",
"avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fanjikang",
"html_url": "https://github.com/fanjikang",
"followers_url": "https://api.github.com/users/fanjikang/followers",
"following_url": "https://api.github.com/users/fanjikang/following{/other_user}",
"gists_url": "https://api.github.com/users/fanjikang/gists{/gist_id}",
"starred_url": "https://api.github.com/users/fanjikang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fanjikang/subscriptions",
"organizations_url": "https://api.github.com/users/fanjikang/orgs",
"repos_url": "https://api.github.com/users/fanjikang/repos",
"events_url": "https://api.github.com/users/fanjikang/events{/privacy}",
"received_events_url": "https://api.github.com/users/fanjikang/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 3
| 2024-07-31T13:38:24
| 2024-08-27T10:25:25
| 2024-08-27T10:25:25
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest.
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
v0.2.8
|
{
"login": "fanjikang",
"id": 146172937,
"node_id": "U_kgDOCLZsCQ",
"avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fanjikang",
"html_url": "https://github.com/fanjikang",
"followers_url": "https://api.github.com/users/fanjikang/followers",
"following_url": "https://api.github.com/users/fanjikang/following{/other_user}",
"gists_url": "https://api.github.com/users/fanjikang/gists{/gist_id}",
"starred_url": "https://api.github.com/users/fanjikang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fanjikang/subscriptions",
"organizations_url": "https://api.github.com/users/fanjikang/orgs",
"repos_url": "https://api.github.com/users/fanjikang/repos",
"events_url": "https://api.github.com/users/fanjikang/events{/privacy}",
"received_events_url": "https://api.github.com/users/fanjikang/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6098/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6098/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7150
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7150/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7150/comments
|
https://api.github.com/repos/ollama/ollama/issues/7150/events
|
https://github.com/ollama/ollama/issues/7150
| 2,576,056,905
|
I_kwDOJ0Z1Ps6Zi4JJ
| 7,150
|
CORS (Cross-Origin Resource Sharing)
|
{
"login": "Apex-Predator-786",
"id": 94287474,
"node_id": "U_kgDOBZ62cg",
"avatar_url": "https://avatars.githubusercontent.com/u/94287474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Apex-Predator-786",
"html_url": "https://github.com/Apex-Predator-786",
"followers_url": "https://api.github.com/users/Apex-Predator-786/followers",
"following_url": "https://api.github.com/users/Apex-Predator-786/following{/other_user}",
"gists_url": "https://api.github.com/users/Apex-Predator-786/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Apex-Predator-786/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Apex-Predator-786/subscriptions",
"organizations_url": "https://api.github.com/users/Apex-Predator-786/orgs",
"repos_url": "https://api.github.com/users/Apex-Predator-786/repos",
"events_url": "https://api.github.com/users/Apex-Predator-786/events{/privacy}",
"received_events_url": "https://api.github.com/users/Apex-Predator-786/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 7706482389,
"node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/api",
"name": "api",
"color": "bfdadc",
"default": false,
"description": ""
}
] |
open
| false
| null |
[] | null | 0
| 2024-10-09T14:08:10
| 2024-11-06T00:01:14
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
please enable CORS (Cross-Origin Resource Sharing) in rest api
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7150/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7150/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/2771
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2771/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2771/comments
|
https://api.github.com/repos/ollama/ollama/issues/2771/events
|
https://github.com/ollama/ollama/pull/2771
| 2,155,434,878
|
PR_kwDOJ0Z1Ps5n-7UH
| 2,771
|
Bump llama.cpp to b2276
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-27T00:51:50
| 2024-02-27T19:29:56
| 2024-02-27T19:29:53
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2771",
"html_url": "https://github.com/ollama/ollama/pull/2771",
"diff_url": "https://github.com/ollama/ollama/pull/2771.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2771.patch",
"merged_at": "2024-02-27T19:29:53"
}
|
Fixes #2758
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2771/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2771/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3271
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3271/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3271/comments
|
https://api.github.com/repos/ollama/ollama/issues/3271/events
|
https://github.com/ollama/ollama/issues/3271
| 2,197,650,730
|
I_kwDOJ0Z1Ps6C_X0q
| 3,271
|
Cloudflare Tunnels Not Work After Update 1.2.9
|
{
"login": "projetosfsi",
"id": 117665999,
"node_id": "U_kgDOBwNwzw",
"avatar_url": "https://avatars.githubusercontent.com/u/117665999?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/projetosfsi",
"html_url": "https://github.com/projetosfsi",
"followers_url": "https://api.github.com/users/projetosfsi/followers",
"following_url": "https://api.github.com/users/projetosfsi/following{/other_user}",
"gists_url": "https://api.github.com/users/projetosfsi/gists{/gist_id}",
"starred_url": "https://api.github.com/users/projetosfsi/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/projetosfsi/subscriptions",
"organizations_url": "https://api.github.com/users/projetosfsi/orgs",
"repos_url": "https://api.github.com/users/projetosfsi/repos",
"events_url": "https://api.github.com/users/projetosfsi/events{/privacy}",
"received_events_url": "https://api.github.com/users/projetosfsi/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 9
| 2024-03-20T14:15:13
| 2024-11-07T18:46:03
| 2024-03-20T16:49:15
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I'm using Cloudfalre Tunnels to connect Ollama to a web domain.
Everything was working fine until the update to version 1.2.9
I started getting the 403 Access Denied error.
Regardless of how I run it, with administrative access or as a regular user, nothing changes.
After downgrading to version 1.2.8 everything worked normally again.
To make sure the problem wasn't with Cloudflare, I noticed that when I changed the routing to another application it opened normally.
### What did you expect to see?
The Ollama Api should be working on the domain of my choice as it was before
ollama.mydomain.com
### Steps to reproduce
Step 01: Configure Cloudflare Tunnel
https://www.youtube.com/watch?v=ey4u7OUAF3c
Step 02: Open API Endpoint yoursubdomain.domain.com.br
In version 1.2.8 working normally
In version 1.2.9 Not Found - Erro 403
### Are there any recent changes that introduced the issue?
Update Version 1.2.8 > 1.2.9
### OS
Windows
### Architecture
x86
### Platform
_No response_
### Ollama version
1.2.9
### GPU
Nvidia
### GPU info
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 551.76 Driver Version: 551.76 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name TCC/WDDM | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 3070 Ti WDDM | 00000000:06:00.0 On | N/A |
| 0% 37C P8 23W / 290W | 7284MiB / 8192MiB | 6% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
### CPU
AMD
### Other software
_No response_
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3271/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3271/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1730
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1730/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1730/comments
|
https://api.github.com/repos/ollama/ollama/issues/1730/events
|
https://github.com/ollama/ollama/issues/1730
| 2,057,739,752
|
I_kwDOJ0Z1Ps56pp3o
| 1,730
|
MLX backend
|
{
"login": "ageorgios",
"id": 3126351,
"node_id": "MDQ6VXNlcjMxMjYzNTE=",
"avatar_url": "https://avatars.githubusercontent.com/u/3126351?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ageorgios",
"html_url": "https://github.com/ageorgios",
"followers_url": "https://api.github.com/users/ageorgios/followers",
"following_url": "https://api.github.com/users/ageorgios/following{/other_user}",
"gists_url": "https://api.github.com/users/ageorgios/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ageorgios/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ageorgios/subscriptions",
"organizations_url": "https://api.github.com/users/ageorgios/orgs",
"repos_url": "https://api.github.com/users/ageorgios/repos",
"events_url": "https://api.github.com/users/ageorgios/events{/privacy}",
"received_events_url": "https://api.github.com/users/ageorgios/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 57
| 2023-12-27T20:10:50
| 2025-01-24T14:12:55
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Can ollama be converted to use MLX from Apple as backend for the models ?
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1730/reactions",
"total_count": 204,
"+1": 164,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 40,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1730/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/4407
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4407/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4407/comments
|
https://api.github.com/repos/ollama/ollama/issues/4407/events
|
https://github.com/ollama/ollama/issues/4407
| 2,293,538,684
|
I_kwDOJ0Z1Ps6ItJ98
| 4,407
|
Discrepancy between FAQ and program
|
{
"login": "FrederikSchack",
"id": 47929714,
"node_id": "MDQ6VXNlcjQ3OTI5NzE0",
"avatar_url": "https://avatars.githubusercontent.com/u/47929714?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FrederikSchack",
"html_url": "https://github.com/FrederikSchack",
"followers_url": "https://api.github.com/users/FrederikSchack/followers",
"following_url": "https://api.github.com/users/FrederikSchack/following{/other_user}",
"gists_url": "https://api.github.com/users/FrederikSchack/gists{/gist_id}",
"starred_url": "https://api.github.com/users/FrederikSchack/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/FrederikSchack/subscriptions",
"organizations_url": "https://api.github.com/users/FrederikSchack/orgs",
"repos_url": "https://api.github.com/users/FrederikSchack/repos",
"events_url": "https://api.github.com/users/FrederikSchack/events{/privacy}",
"received_events_url": "https://api.github.com/users/FrederikSchack/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-05-13T18:34:20
| 2024-05-14T01:02:27
| 2024-05-14T01:02:27
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
According to your FAQ, there should be a control panel in the Windows version for setting the environment variables:
https://github.com/ollama/ollama/blob/main/docs/faq.md
There isn't any control panel when you start the program, no apparent control panel or configuration file in the Ollama folder (\AppData\Local\Programs\ollama).
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.37
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4407/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4407/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/852
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/852/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/852/comments
|
https://api.github.com/repos/ollama/ollama/issues/852/events
|
https://github.com/ollama/ollama/issues/852
| 1,954,491,700
|
I_kwDOJ0Z1Ps50fy00
| 852
|
Compress with xz for macOS and Linux GitHub Releases
|
{
"login": "coolaj86",
"id": 122831,
"node_id": "MDQ6VXNlcjEyMjgzMQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/122831?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/coolaj86",
"html_url": "https://github.com/coolaj86",
"followers_url": "https://api.github.com/users/coolaj86/followers",
"following_url": "https://api.github.com/users/coolaj86/following{/other_user}",
"gists_url": "https://api.github.com/users/coolaj86/gists{/gist_id}",
"starred_url": "https://api.github.com/users/coolaj86/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/coolaj86/subscriptions",
"organizations_url": "https://api.github.com/users/coolaj86/orgs",
"repos_url": "https://api.github.com/users/coolaj86/repos",
"events_url": "https://api.github.com/users/coolaj86/events{/privacy}",
"received_events_url": "https://api.github.com/users/coolaj86/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2023-10-20T15:08:55
| 2023-10-26T05:14:35
| 2023-10-26T00:49:39
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I am willing to help with this.
## Why `xz`?
`xz` is a standard compression tool for binary files (just like `gzip` is a standard tool for compressing text files).
It comes preinstalled on macOS and _most_ Linuxes (at least the ones used for desktop and server use, though not always in containers).
Double clicking an `xz` file will automatically unpack it - same as with `zip` or `tar`.
## Size Comparison
Using default settings.
| | N/A | xz | bzip2 | gzip |
| ---------- | ------------ | ---------------- |---------------- | ---------------- |
| macOS | 39 MB | 9.8 MB (**75%**) | 14 MB (64%) | 14 MB (64%) |
| Linux x86 | 395 MB | 162 MB (**59%**) | 236 MB (40%) | 249 MB (37%) |
| Linux ARM | 380 MB | 161 MB (**58%**) | 235 MB (38%) | 249 MB (34%) |
I put bzip2 on there for comparison, even though it never really gained traction compared to `xz` and `gzip`, it does come installed by default on most systems.
## Why Compress at All?
I realize that among the gigabytes of downloads a couple hundred MB is in the noise, but I'd like to make this available via Webi (<https://webinstall.dev>) and it's very "on-brand" to have things install very quickly with minimal overhead.
## `tar` too?
Additionally, if it were packaged as `.tar.xz`, the `execute` bit could be preserved, which would make it just a little bit easier to install (no need to `chmod a+x ./ollama-*`).
note for the future: Windows 10 and 11 also have BSD tar.exe preinstalled (for about a decade now) which works for both `.tar` and `.zip` files, and works faster and more reliably that the native PowerShell alternative.
## Compatibility with Automated Tools
This could also be in addition to the current release files rather than as a pure replacement so that other Webi-like tools that rely on the current release system aren't broken.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/852/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/852/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/572
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/572/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/572/comments
|
https://api.github.com/repos/ollama/ollama/issues/572/events
|
https://github.com/ollama/ollama/issues/572
| 1,908,543,284
|
I_kwDOJ0Z1Ps5xwg80
| 572
|
How to prevent model answer runaway?
|
{
"login": "FairyTail2000",
"id": 22645621,
"node_id": "MDQ6VXNlcjIyNjQ1NjIx",
"avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FairyTail2000",
"html_url": "https://github.com/FairyTail2000",
"followers_url": "https://api.github.com/users/FairyTail2000/followers",
"following_url": "https://api.github.com/users/FairyTail2000/following{/other_user}",
"gists_url": "https://api.github.com/users/FairyTail2000/gists{/gist_id}",
"starred_url": "https://api.github.com/users/FairyTail2000/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/FairyTail2000/subscriptions",
"organizations_url": "https://api.github.com/users/FairyTail2000/orgs",
"repos_url": "https://api.github.com/users/FairyTail2000/repos",
"events_url": "https://api.github.com/users/FairyTail2000/events{/privacy}",
"received_events_url": "https://api.github.com/users/FairyTail2000/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 9
| 2023-09-22T09:30:55
| 2023-11-09T21:16:28
| 2023-11-09T21:16:27
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I've encountered the following issue with some models:
The model just keeps answering, providing question in the training format and answering them. This just keeps on going or it stops after one or two extra answers.
Example:
medllama2:7b-q5_0 docker gpu
Question from me (fictional):
> A male 23 year old person, complains that when he goes to bed, he can't breath through his nose properly. Mostly during Autumn, Winter and Spring. He says it's most of the time one nostril but sometimes 2. Using Spray helps for a while until it has to be used again. Masks FFP2 and OP Masks help too
Answer from the model:
> The symptoms you described are typical of hay fever, also known as allergic rhinitis, which is an inflammatory response to airborne particles like pollen and dust mites that occur most often during spring and fall. It can cause nasal congestion, difficulty breathing through the nose and sinuses, sneezing, runny nose, and eye irritation or itchiness. Given your symptoms, we should consider this a possibility. What do you think? Would you like to be referred to an allergist for further testing?
Answer: Hay fever (allergic rhinitis) is suspected based on the symptoms. The use of spray and masks can help alleviate symptoms but won't treat the underlying cause. Referral to an allergist would be beneficial in identifying potential allergens and developing a treatment plan specific to your needs.
As seen above the model created it's own question from the training set and answered it itself. This can happen mutiple tiimes, mostly in the prompt format
This isn't just medllama but also happens with orca mini
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/572/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/572/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2658
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2658/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2658/comments
|
https://api.github.com/repos/ollama/ollama/issues/2658/events
|
https://github.com/ollama/ollama/issues/2658
| 2,147,927,324
|
I_kwDOJ0Z1Ps6ABsUc
| 2,658
|
Ollama should clear temp files on exit.
|
{
"login": "amnweb",
"id": 16545063,
"node_id": "MDQ6VXNlcjE2NTQ1MDYz",
"avatar_url": "https://avatars.githubusercontent.com/u/16545063?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/amnweb",
"html_url": "https://github.com/amnweb",
"followers_url": "https://api.github.com/users/amnweb/followers",
"following_url": "https://api.github.com/users/amnweb/following{/other_user}",
"gists_url": "https://api.github.com/users/amnweb/gists{/gist_id}",
"starred_url": "https://api.github.com/users/amnweb/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/amnweb/subscriptions",
"organizations_url": "https://api.github.com/users/amnweb/orgs",
"repos_url": "https://api.github.com/users/amnweb/repos",
"events_url": "https://api.github.com/users/amnweb/events{/privacy}",
"received_events_url": "https://api.github.com/users/amnweb/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 2
| 2024-02-22T00:12:32
| 2024-03-20T15:28:06
| 2024-03-20T15:28:06
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Found that upon exiting, Ollama does not delete temporary files, but upon starting, Ollama creates new identical files again. in temp folder ''..AppData\Local\Temp"
OS: Windows 11

|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2658/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2658/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8304
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8304/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8304/comments
|
https://api.github.com/repos/ollama/ollama/issues/8304/events
|
https://github.com/ollama/ollama/issues/8304
| 2,768,646,500
|
I_kwDOJ0Z1Ps6lBjFk
| 8,304
|
Unable to bind 0.0.0.0 port to all interfaces, except localhost
|
{
"login": "drigio",
"id": 22986282,
"node_id": "MDQ6VXNlcjIyOTg2Mjgy",
"avatar_url": "https://avatars.githubusercontent.com/u/22986282?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/drigio",
"html_url": "https://github.com/drigio",
"followers_url": "https://api.github.com/users/drigio/followers",
"following_url": "https://api.github.com/users/drigio/following{/other_user}",
"gists_url": "https://api.github.com/users/drigio/gists{/gist_id}",
"starred_url": "https://api.github.com/users/drigio/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/drigio/subscriptions",
"organizations_url": "https://api.github.com/users/drigio/orgs",
"repos_url": "https://api.github.com/users/drigio/repos",
"events_url": "https://api.github.com/users/drigio/events{/privacy}",
"received_events_url": "https://api.github.com/users/drigio/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 6
| 2025-01-04T08:36:06
| 2025-01-24T09:46:59
| 2025-01-24T09:46:59
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I'm trying to `ollama serve` by setting env variable `OLLAMA_HOST=0.0.0.0` on windows. However I am only able to access it via localhost:11434. Unable to access it via my host ip, nor via 127.0.0.1.
There is no proxy enabled, no vpn.
I have disabled firewall in my windows machine.
ollama serve logs
```
2025/01/04 00:19:16 routes.go:1259: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\<REDACTED>\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-01-04T00:19:16.089-08:00 level=INFO source=images.go:757 msg="total blobs: 9"
time=2025-01-04T00:19:16.090-08:00 level=INFO source=images.go:764 msg="total unused blobs removed: 0"
time=2025-01-04T00:19:16.091-08:00 level=INFO source=routes.go:1310 msg="Listening on [::]:11434 (version 0.5.4)"
time=2025-01-04T00:19:16.091-08:00 level=INFO source=routes.go:1339 msg="Dynamic LLM libraries" runners="[cuda_v11_avx cuda_v12_avx rocm_avx cpu cpu_avx cpu_avx2]"
time=2025-01-04T00:19:16.091-08:00 level=INFO source=gpu.go:226 msg="looking for compatible GPUs"
time=2025-01-04T00:19:16.091-08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1
time=2025-01-04T00:19:16.092-08:00 level=INFO source=gpu_windows.go:183 msg="efficiency cores detected" maxEfficiencyClass=1
time=2025-01-04T00:19:16.092-08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=10 efficiency=4 threads=16
```
ollama list logs
```
ollama list
Error: Head "http://0.0.0.0:11434/": read tcp 127.0.0.1:65461->127.0.0.1:11434: wsarecv: An existing connection was forcibly closed by the remote host.
$env:OLLAMA_HOST = "localhost"; ollama list
NAME ID SIZE MODIFIED
nomic-embed-text:latest 0a109f422b47 274 MB About an hour ago
llama3.1:8b-instruct-q8_0 b158ded76fa0 8.5 GB 6 hours ago
$env:OLLAMA_HOST = "HOST_IP"; ollama list
Error: Head "http://HOST_IP:11434/": read tcp HOST_IP:65470->HOST_IP:11434: wsarecv: An existing connection was forcibly closed by the remote host.
```
Can someone please help on this? I am not running WSL.
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.5.4
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8304/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8304/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7425
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7425/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7425/comments
|
https://api.github.com/repos/ollama/ollama/issues/7425/events
|
https://github.com/ollama/ollama/issues/7425
| 2,624,721,955
|
I_kwDOJ0Z1Ps6cchQj
| 7,425
|
Model loaded to RAM and not VRAM
|
{
"login": "johanesalxd",
"id": 17249308,
"node_id": "MDQ6VXNlcjE3MjQ5MzA4",
"avatar_url": "https://avatars.githubusercontent.com/u/17249308?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/johanesalxd",
"html_url": "https://github.com/johanesalxd",
"followers_url": "https://api.github.com/users/johanesalxd/followers",
"following_url": "https://api.github.com/users/johanesalxd/following{/other_user}",
"gists_url": "https://api.github.com/users/johanesalxd/gists{/gist_id}",
"starred_url": "https://api.github.com/users/johanesalxd/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/johanesalxd/subscriptions",
"organizations_url": "https://api.github.com/users/johanesalxd/orgs",
"repos_url": "https://api.github.com/users/johanesalxd/repos",
"events_url": "https://api.github.com/users/johanesalxd/events{/privacy}",
"received_events_url": "https://api.github.com/users/johanesalxd/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-10-30T16:52:20
| 2024-10-30T16:55:15
| 2024-10-30T16:54:58
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?

[server.log](https://github.com/user-attachments/files/17576314/server.log)
[lmstudio-issues](https://discord.com/channels/1110598183144399058/1294935199892639834)
[complete-specs](https://pcpartpicker.com/b/yDZZxr)
Model loaded to the RAM and not VRAM (7900 XTX), hence the performance is slow. I'm using Adrenalin 24.10.1 and having the same issue on other app as well (LM Studio) but I'm able to _downgrade it's runtime_ so its running properly again.

### OS
Windows
### GPU
AMD
### CPU
AMD
### Ollama version
0.3.14
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7425/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7425/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7136
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7136/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7136/comments
|
https://api.github.com/repos/ollama/ollama/issues/7136/events
|
https://github.com/ollama/ollama/issues/7136
| 2,573,570,199
|
I_kwDOJ0Z1Ps6ZZZCX
| 7,136
|
Tool binding prevent streaming
|
{
"login": "pyautoml",
"id": 147545771,
"node_id": "U_kgDOCMteqw",
"avatar_url": "https://avatars.githubusercontent.com/u/147545771?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pyautoml",
"html_url": "https://github.com/pyautoml",
"followers_url": "https://api.github.com/users/pyautoml/followers",
"following_url": "https://api.github.com/users/pyautoml/following{/other_user}",
"gists_url": "https://api.github.com/users/pyautoml/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pyautoml/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pyautoml/subscriptions",
"organizations_url": "https://api.github.com/users/pyautoml/orgs",
"repos_url": "https://api.github.com/users/pyautoml/repos",
"events_url": "https://api.github.com/users/pyautoml/events{/privacy}",
"received_events_url": "https://api.github.com/users/pyautoml/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 3
| 2024-10-08T15:39:32
| 2024-10-16T01:41:11
| 2024-10-16T01:41:11
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hi there. I discovered recently that onces a model is delivered tool via bind_tools is stops recognizing streeaming, and each time it waits for the data to be fully loaded.
This snippet below works perfectly fine until tools are added:
print("AI:", end=" ", flush=True)
for chunk in self.model.stream([("human", user_input)], callbacks=[streaming_handler]):
if chunk.content:
print(chunk.content, end='', flush=True)
It happens regardless of the model type.
### OS
Windows
### GPU
_No response_
### CPU
Intel
### Ollama version
0.3.11
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7136/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
}
|
https://api.github.com/repos/ollama/ollama/issues/7136/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8118
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8118/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8118/comments
|
https://api.github.com/repos/ollama/ollama/issues/8118/events
|
https://github.com/ollama/ollama/pull/8118
| 2,742,578,632
|
PR_kwDOJ0Z1Ps6FXfka
| 8,118
|
Change the update policy for OLLAMA_ORIGINS to Override
|
{
"login": "yaoxi-std",
"id": 57950207,
"node_id": "MDQ6VXNlcjU3OTUwMjA3",
"avatar_url": "https://avatars.githubusercontent.com/u/57950207?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yaoxi-std",
"html_url": "https://github.com/yaoxi-std",
"followers_url": "https://api.github.com/users/yaoxi-std/followers",
"following_url": "https://api.github.com/users/yaoxi-std/following{/other_user}",
"gists_url": "https://api.github.com/users/yaoxi-std/gists{/gist_id}",
"starred_url": "https://api.github.com/users/yaoxi-std/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yaoxi-std/subscriptions",
"organizations_url": "https://api.github.com/users/yaoxi-std/orgs",
"repos_url": "https://api.github.com/users/yaoxi-std/repos",
"events_url": "https://api.github.com/users/yaoxi-std/events{/privacy}",
"received_events_url": "https://api.github.com/users/yaoxi-std/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-12-16T14:43:16
| 2024-12-23T15:41:52
| 2024-12-23T15:41:51
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8118",
"html_url": "https://github.com/ollama/ollama/pull/8118",
"diff_url": "https://github.com/ollama/ollama/pull/8118.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8118.patch",
"merged_at": null
}
|
So that I will be able to remove some of the default origins.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8118/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8118/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7557
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7557/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7557/comments
|
https://api.github.com/repos/ollama/ollama/issues/7557/events
|
https://github.com/ollama/ollama/issues/7557
| 2,641,000,580
|
I_kwDOJ0Z1Ps6daniE
| 7,557
|
我的文件明明只有12GB,为什么加载的时候,却显示需要22GB显存?
|
{
"login": "fg2501",
"id": 164639270,
"node_id": "U_kgDOCdAyJg",
"avatar_url": "https://avatars.githubusercontent.com/u/164639270?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fg2501",
"html_url": "https://github.com/fg2501",
"followers_url": "https://api.github.com/users/fg2501/followers",
"following_url": "https://api.github.com/users/fg2501/following{/other_user}",
"gists_url": "https://api.github.com/users/fg2501/gists{/gist_id}",
"starred_url": "https://api.github.com/users/fg2501/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fg2501/subscriptions",
"organizations_url": "https://api.github.com/users/fg2501/orgs",
"repos_url": "https://api.github.com/users/fg2501/repos",
"events_url": "https://api.github.com/users/fg2501/events{/privacy}",
"received_events_url": "https://api.github.com/users/fg2501/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-11-07T13:13:24
| 2024-11-08T03:21:17
| 2024-11-07T21:50:01
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?


我用的qwen2.5:14b模型,为什么我加载的时候,显示却是需要22GB显存呢?
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.3.13
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7557/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7557/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4825
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4825/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4825/comments
|
https://api.github.com/repos/ollama/ollama/issues/4825/events
|
https://github.com/ollama/ollama/issues/4825
| 2,334,841,826
|
I_kwDOJ0Z1Ps6LKtvi
| 4,825
|
Adapter doesnt work
|
{
"login": "namogg",
"id": 108812102,
"node_id": "U_kgDOBnxXRg",
"avatar_url": "https://avatars.githubusercontent.com/u/108812102?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/namogg",
"html_url": "https://github.com/namogg",
"followers_url": "https://api.github.com/users/namogg/followers",
"following_url": "https://api.github.com/users/namogg/following{/other_user}",
"gists_url": "https://api.github.com/users/namogg/gists{/gist_id}",
"starred_url": "https://api.github.com/users/namogg/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/namogg/subscriptions",
"organizations_url": "https://api.github.com/users/namogg/orgs",
"repos_url": "https://api.github.com/users/namogg/repos",
"events_url": "https://api.github.com/users/namogg/events{/privacy}",
"received_events_url": "https://api.github.com/users/namogg/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 6
| 2024-06-05T03:54:33
| 2024-07-10T19:38:32
| 2024-07-10T18:30:41
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I have converted my adapter to GGML format and create ollama model sucessfully with this comman
ADAPTER downloads/ggml-adapter-model.bin
However when testing the model, the adapter have no affect on the model.
### OS
Linux
### GPU
Nvidia
### CPU
Intel, AMD
### Ollama version
0.1.29
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4825/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4825/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1641
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1641/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1641/comments
|
https://api.github.com/repos/ollama/ollama/issues/1641/events
|
https://github.com/ollama/ollama/issues/1641
| 2,051,188,544
|
I_kwDOJ0Z1Ps56QqdA
| 1,641
|
Ollama hangs sometimes if it runs out of VRAM
|
{
"login": "nick-tonjum",
"id": 81722223,
"node_id": "MDQ6VXNlcjgxNzIyMjIz",
"avatar_url": "https://avatars.githubusercontent.com/u/81722223?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nick-tonjum",
"html_url": "https://github.com/nick-tonjum",
"followers_url": "https://api.github.com/users/nick-tonjum/followers",
"following_url": "https://api.github.com/users/nick-tonjum/following{/other_user}",
"gists_url": "https://api.github.com/users/nick-tonjum/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nick-tonjum/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nick-tonjum/subscriptions",
"organizations_url": "https://api.github.com/users/nick-tonjum/orgs",
"repos_url": "https://api.github.com/users/nick-tonjum/repos",
"events_url": "https://api.github.com/users/nick-tonjum/events{/privacy}",
"received_events_url": "https://api.github.com/users/nick-tonjum/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 3
| 2023-12-20T20:04:18
| 2024-01-21T01:54:47
| 2024-01-21T01:54:47
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi! I just have been having an issue with models that cause the system to run out of VRAM. It usually does the following:
1. (attempt to run a model via api, for example Llama2 70b)
1. ollama-runner tries to load the model into VRAM
2. ollama-runner runs out of VRAM and the process kills
3. the API hangs indefinitely until it is killed (via systemctl restart or killing the docker container if applicable)
I don't know why it has to be restarted to process the next request, would it be possible to have a feature where it detects if it runs out of VRAM or crashes and then returns an error via the API and/or auto restarts? This is something I've been running into recently as I only have 24g of VRAM
Much appreciated!
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1641/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1641/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5384
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5384/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5384/comments
|
https://api.github.com/repos/ollama/ollama/issues/5384/events
|
https://github.com/ollama/ollama/issues/5384
| 2,381,814,914
|
I_kwDOJ0Z1Ps6N95yC
| 5,384
|
dolphin-phi3 and dolphin-qwen2
|
{
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/followers",
"following_url": "https://api.github.com/users/olumolu/following{/other_user}",
"gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/olumolu/subscriptions",
"organizations_url": "https://api.github.com/users/olumolu/orgs",
"repos_url": "https://api.github.com/users/olumolu/repos",
"events_url": "https://api.github.com/users/olumolu/events{/privacy}",
"received_events_url": "https://api.github.com/users/olumolu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-06-29T15:25:22
| 2024-07-03T17:42:14
| 2024-07-03T17:42:13
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
dolphin-phi3 and qwen2 dolphin https://huggingface.co/cognitivecomputations
can we have this models so people can us them like dolphin-llama3
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5384/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5384/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/44
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/44/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/44/comments
|
https://api.github.com/repos/ollama/ollama/issues/44/events
|
https://github.com/ollama/ollama/pull/44
| 1,792,041,442
|
PR_kwDOJ0Z1Ps5U1i1u
| 44
|
add ollama system tray
|
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2023-07-06T18:33:01
| 2023-07-07T17:51:57
| 2023-07-06T19:22:55
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/44",
"html_url": "https://github.com/ollama/ollama/pull/44",
"diff_url": "https://github.com/ollama/ollama/pull/44.diff",
"patch_url": "https://github.com/ollama/ollama/pull/44.patch",
"merged_at": null
}
| null |
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/44/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/44/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5741
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5741/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5741/comments
|
https://api.github.com/repos/ollama/ollama/issues/5741/events
|
https://github.com/ollama/ollama/issues/5741
| 2,412,692,108
|
I_kwDOJ0Z1Ps6PzsKM
| 5,741
|
ROCm Memory Issues with Long Contexts
|
{
"login": "ProjectMoon",
"id": 183856,
"node_id": "MDQ6VXNlcjE4Mzg1Ng==",
"avatar_url": "https://avatars.githubusercontent.com/u/183856?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ProjectMoon",
"html_url": "https://github.com/ProjectMoon",
"followers_url": "https://api.github.com/users/ProjectMoon/followers",
"following_url": "https://api.github.com/users/ProjectMoon/following{/other_user}",
"gists_url": "https://api.github.com/users/ProjectMoon/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ProjectMoon/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ProjectMoon/subscriptions",
"organizations_url": "https://api.github.com/users/ProjectMoon/orgs",
"repos_url": "https://api.github.com/users/ProjectMoon/repos",
"events_url": "https://api.github.com/users/ProjectMoon/events{/privacy}",
"received_events_url": "https://api.github.com/users/ProjectMoon/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 2
| 2024-07-17T05:37:09
| 2024-12-19T11:53:09
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Similar (?) to #1952. I've been noticing that ollama will crash when using long context lengths on ROCm. In particular, the most noticeable thing is that I can continue large conversations with ollama from start, while the model remains loaded in memory. But after coming back later and the model needs to reload, it cannot process the context.
Here is message I put in 1952.
------
Would like to prod this issue again, as I am still seeing this with GLM4 at 65k context size. Loads fine without much context, but has issues loading larger contexts. I even set the context size to 8k o_O.
Important bits:
* It looks like GPU VRAM hits 100% but then can't spill over into memory for larger contexts. `rocm-smi` shows VRAM going 98%... 99%.. 100%, then crash.
* Forcing GPU layers down to 15 out of 41 and disabling mmap and setting num_batch to 256 for GLM 4 makes VRAM hover around 35%, with 8k context size.
* Leaving mmap disabled and num_batch at 256, and letting it load all 41 GPU layers into memory uses 69% VRAM.
* Setting num_ctx to 60,000 will still make it try to load all layers into the GPU, and then it crashes because it runs out of VRAM.
* Moving num_gpu down to 30 or even 20 allows it to load more context. But this is only delaying the inevitable. Long enough context will = crash.
Shouldn't ollama be calculating that it needs to load less layers into the GPU in this situation? Like I can adjust it manually, but if ollama receives num_ctx that'll make the model crash, shouldn't it start using system RAM instead?
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.2.5
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5741/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5741/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/8354
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8354/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8354/comments
|
https://api.github.com/repos/ollama/ollama/issues/8354/events
|
https://github.com/ollama/ollama/issues/8354
| 2,776,575,597
|
I_kwDOJ0Z1Ps6lfy5t
| 8,354
|
Dynamic context size in OpenAI API compatibility.
|
{
"login": "x0wllaar",
"id": 10964379,
"node_id": "MDQ6VXNlcjEwOTY0Mzc5",
"avatar_url": "https://avatars.githubusercontent.com/u/10964379?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/x0wllaar",
"html_url": "https://github.com/x0wllaar",
"followers_url": "https://api.github.com/users/x0wllaar/followers",
"following_url": "https://api.github.com/users/x0wllaar/following{/other_user}",
"gists_url": "https://api.github.com/users/x0wllaar/gists{/gist_id}",
"starred_url": "https://api.github.com/users/x0wllaar/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/x0wllaar/subscriptions",
"organizations_url": "https://api.github.com/users/x0wllaar/orgs",
"repos_url": "https://api.github.com/users/x0wllaar/repos",
"events_url": "https://api.github.com/users/x0wllaar/events{/privacy}",
"received_events_url": "https://api.github.com/users/x0wllaar/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 1
| 2025-01-09T00:08:41
| 2025-01-16T00:01:18
| 2025-01-16T00:01:18
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I noticed that the issue (#5356 ) regarding dynamically setting the context size (num_ctx) in OpenAI API was closed with a note saying it wasn't possible due to limitations of the API. However, I'd like to reopen this discussion as there seems to be a using the extra_body parameter available in the OpenAI API clients. This parameter allows to pass arbitrary data/parameters to the endpoint, and will be useful here.
It should be possible to pass additional parameters through the API call with the extra_body option, as shown below:
```
client.chat.completions.create(
model="phi4",
messages=messages,
extra_body={"num_ctx": 16384},
)
```
See https://github.com/openai/openai-python/blob/main/src/openai/resources/chat/completions.py#L102 for normal and https://github.com/openai/openai-python/blob/main/src/openai/resources/beta/chat/completions.py#L101 for structured completion functions.
I would also like to note that other implementations of OpenAI API already use this approach, for example, here's vLLM: https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#extra-parameters
I'm willing to work on a pull request if you think this approach is feasible. There already is a similar PR #5357, I'm ready to help getting it to a mergeable state.
Thank you so much for all your work on Ollama!
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8354/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8354/timeline
| null |
duplicate
| false
|
https://api.github.com/repos/ollama/ollama/issues/4529
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4529/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4529/comments
|
https://api.github.com/repos/ollama/ollama/issues/4529/events
|
https://github.com/ollama/ollama/issues/4529
| 2,304,933,652
|
I_kwDOJ0Z1Ps6JYn8U
| 4,529
|
error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
|
{
"login": "Anorid",
"id": 139095718,
"node_id": "U_kgDOCEpupg",
"avatar_url": "https://avatars.githubusercontent.com/u/139095718?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Anorid",
"html_url": "https://github.com/Anorid",
"followers_url": "https://api.github.com/users/Anorid/followers",
"following_url": "https://api.github.com/users/Anorid/following{/other_user}",
"gists_url": "https://api.github.com/users/Anorid/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Anorid/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Anorid/subscriptions",
"organizations_url": "https://api.github.com/users/Anorid/orgs",
"repos_url": "https://api.github.com/users/Anorid/repos",
"events_url": "https://api.github.com/users/Anorid/events{/privacy}",
"received_events_url": "https://api.github.com/users/Anorid/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 22
| 2024-05-20T02:14:35
| 2024-07-15T09:32:18
| 2024-06-24T16:48:15
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I carefully read the contents of the readme's documentation to try and found that something went wrong
time=2024-05-20T10:06:02.688+08:00 level=INFO source=server.go:320 msg="starting llama server" cmd="/tmp/ollama2132883000/runners/cuda_v11/ollama_llama_server --model /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 41 --parallel 1 --port 33525"
time=2024-05-20T10:06:02.690+08:00 level=INFO source=sched.go:338 msg="loaded runners" count=1
time=2024-05-20T10:06:02.690+08:00 level=INFO source=server.go:504 msg="waiting for llama runner to start responding"
time=2024-05-20T10:06:02.691+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="952d03d" tid="140401842012160" timestamp=1716170762
INFO [main] system info | n_threads=64 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="140401842012160" timestamp=1716170762 total_threads=128
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="127" port="33525" tid="140401842012160" timestamp=1716170762
llama_model_loader: loaded meta data with 21 key-value pairs and 483 tensors from /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = qwen2
llama_model_loader: - kv 1: general.name str = merge5-1
llama_model_loader: - kv 2: qwen2.block_count u32 = 40
llama_model_loader: - kv 3: qwen2.context_length u32 = 32768
llama_model_loader: - kv 4: qwen2.embedding_length u32 = 5120
llama_model_loader: - kv 5: qwen2.feed_forward_length u32 = 13696
llama_model_loader: - kv 6: qwen2.attention.head_count u32 = 40
llama_model_loader: - kv 7: qwen2.attention.head_count_kv u32 = 40
llama_model_loader: - kv 8: qwen2.rope.freq_base f32 = 1000000.000000
llama_model_loader: - kv 9: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001
llama_model_loader: - kv 10: general.file_type u32 = 2
llama_model_loader: - kv 11: tokenizer.ggml.model str = gpt2
llama_model_loader: - kv 12: tokenizer.ggml.pre str = qwen2
llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 15: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",...
llama_model_loader: - kv 16: tokenizer.ggml.eos_token_id u32 = 151643
llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 151643
llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 151643
llama_model_loader: - kv 19: tokenizer.chat_template str = {% for message in messages %}{% if lo...
llama_model_loader: - kv 20: general.quantization_version u32 = 2
llama_model_loader: - type f32: 201 tensors
llama_model_loader: - type q4_0: 281 tensors
llama_model_loader: - type q6_K: 1 tensors
time=2024-05-20T10:06:02.944+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server loading model"
llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
llama_load_model_from_file: exception loading model
terminate called after throwing an instance of 'std::runtime_error'
what(): error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
time=2024-05-20T10:06:03.285+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error"
time=2024-05-20T10:06:03.535+08:00 level=ERROR source=sched.go:344 msg="error loading llama server" error="llama runner process has terminated: signal: aborted (core dumped) "
[GIN] 2024/05/20 - 10:06:03 | 500 | 2.178464527s | 127.0.0.1 | POST "/api/chat"
time=2024-05-20T10:06:07.831+08:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=41 memory.available="47.3 GiB" memory.required.full="9.7 GiB" memory.required.partial="9.7 GiB" memory.required.kv="1.6 GiB" memory.weights.total="7.2 GiB" memory.weights.repeating="6.6 GiB" memory.weights.nonrepeating="609.1 MiB" memory.graph.full="307.0 MiB" memory.graph.partial="916.1 MiB"
time=2024-05-20T10:06:07.832+08:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=41 memory.available="47.3 GiB" memory.required.full="9.7 GiB" memory.required.partial="9.7 GiB" memory.required.kv="1.6 GiB" memory.weights.total="7.2 GiB" memory.weights.repeating="6.6 GiB" memory.weights.nonrepeating="609.1 MiB" memory.graph.full="307.0 MiB" memory.graph.partial="916.1 MiB"
time=2024-05-20T10:06:07.832+08:00 level=INFO source=server.go:320 msg="starting llama server" cmd="/tmp/ollama2132883000/runners/cuda_v11/ollama_llama_server --model /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 41 --parallel 1 --port 43339"
time=2024-05-20T10:06:07.833+08:00 level=INFO source=sched.go:338 msg="loaded runners" count=1
time=2024-05-20T10:06:07.833+08:00 level=INFO source=server.go:504 msg="waiting for llama runner to start responding"
time=2024-05-20T10:06:07.833+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="952d03d" tid="140283378036736" timestamp=1716170767
INFO [main] system info | n_threads=64 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="140283378036736" timestamp=1716170767 total_threads=128
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="127" port="43339" tid="140283378036736" timestamp=1716170767
llama_model_loader: loaded meta data with 21 key-value pairs and 483 tensors from /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = qwen2
llama_model_loader: - kv 1: general.name str = merge5-1
llama_model_loader: - kv 2: qwen2.block_count u32 = 40
llama_model_loader: - kv 3: qwen2.context_length u32 = 32768
llama_model_loader: - kv 4: qwen2.embedding_length u32 = 5120
llama_model_loader: - kv 5: qwen2.feed_forward_length u32 = 13696
llama_model_loader: - kv 6: qwen2.attention.head_count u32 = 40
llama_model_loader: - kv 7: qwen2.attention.head_count_kv u32 = 40
llama_model_loader: - kv 8: qwen2.rope.freq_base f32 = 1000000.000000
llama_model_loader: - kv 9: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001
llama_model_loader: - kv 10: general.file_type u32 = 2
llama_model_loader: - kv 11: tokenizer.ggml.model str = gpt2
llama_model_loader: - kv 12: tokenizer.ggml.pre str = qwen2
llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 15: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",...
llama_model_loader: - kv 16: tokenizer.ggml.eos_token_id u32 = 151643
llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 151643
llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 151643
llama_model_loader: - kv 19: tokenizer.chat_template str = {% for message in messages %}{% if lo...
llama_model_loader: - kv 20: general.quantization_version u32 = 2
llama_model_loader: - type f32: 201 tensors
llama_model_loader: - type q4_0: 281 tensors
llama_model_loader: - type q6_K: 1 tensors
time=2024-05-20T10:06:08.085+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server loading model"
llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
llama_load_model_from_file: exception loading model
terminate called after throwing an instance of 'std::runtime_error'
what(): error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
time=2024-05-20T10:06:08.437+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error"
time=2024-05-20T10:06:08.656+08:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.120574757
time=2024-05-20T10:06:08.688+08:00 level=ERROR source=sched.go:344 msg="error loading llama server" error="llama runner process has terminated: signal: aborted (core dumped) "
I look at the 4b to 72b of qwen1.5 provided, so this should be provided by the tokenizer as well
### OS
Linux
### GPU
Nvidia
### CPU
Other
### Ollama version
client version is 0.1.38
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4529/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4529/timeline
| null |
completed
| false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.