url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/5614
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5614/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5614/comments
https://api.github.com/repos/ollama/ollama/issues/5614/events
https://github.com/ollama/ollama/pull/5614
2,401,766,631
PR_kwDOJ0Z1Ps51BVsT
5,614
OpenAI: Support Tools
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-10T21:21:37
2024-07-23T23:02:27
2024-07-17T03:52:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5614", "html_url": "https://github.com/ollama/ollama/pull/5614", "diff_url": "https://github.com/ollama/ollama/pull/5614.diff", "patch_url": "https://github.com/ollama/ollama/pull/5614.patch", "merged_at": "2024-07-17T03:52:59" }
``` curl -s 127.0.0.1:11434/v1/chat/completions -d '{ "model": "mike/mistral", "messages": [ { "role": "user", "content": "What'\''s the weather like today in Paris?" }, { "role": "assistant", "tool_calls": [ { "id": "89a1e453-0bce-4de3-a456-c54bed09c520", "type": "function", "function": { "name": "get_current_weather", "arguments": "{\"location\": \"Paris, France\", \"format\": \"celsius\"}" } } ] }, { "role": "tool", "tool_call_id": "89a1e453-0bce-4de3-a456-c54bed09c520", "content": "22" }, { "role": "assistant", "content": "The weather in Paris is 22 degrees celsius." }, { "role": "user", "content": "What'\''s the weather like today in San Francisco and Toronto?" } ], "tools": [ { "type": "function", "function": { "name": "get_current_weather", "description": "Get the current weather", "parameters": { "type": "object", "properties": { "location": { "type": "string", "description": "The city and state, e.g. San Francisco, CA" }, "format": { "type": "string", "enum": [ "celsius", "fahrenheit" ], "description": "The temperature unit to use. Infer this from the users location." } }, "required": [ "location", "format" ] } } } ], "stream": false, "temperature": 0 }' | jq ``` ``` { "id": "chatcmpl-920", "object": "chat.completion", "created": 1721187203, "model": "mike/mistral", "system_fingerprint": "fp_ollama", "choices": [ { "index": 0, "message": { "role": "assistant", "content": "", "tool_calls": [ { "id": "call_q9jtgbw2", "type": "function", "function": { "name": "get_current_weather", "arguments": "{\"format\":\"celsius\",\"location\":\"San Francisco, CA\"}" } }, { "id": "call_9chg85cd", "type": "function", "function": { "name": "get_current_weather", "arguments": "{\"format\":\"celsius\",\"location\":\"Toronto, Canada\"}" } } ] }, "finish_reason": "stop" } ], "usage": { "prompt_tokens": 198, "completion_tokens": 65, "total_tokens": 263 } } ```
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5614/reactions", "total_count": 8, "+1": 0, "-1": 0, "laugh": 0, "hooray": 8, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5614/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/700
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/700/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/700/comments
https://api.github.com/repos/ollama/ollama/issues/700/events
https://github.com/ollama/ollama/pull/700
1,926,883,924
PR_kwDOJ0Z1Ps5b77h7
700
rename server subprocess
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-10-04T19:53:55
2023-10-06T14:15:43
2023-10-06T14:15:42
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/700", "html_url": "https://github.com/ollama/ollama/pull/700", "diff_url": "https://github.com/ollama/ollama/pull/700.diff", "patch_url": "https://github.com/ollama/ollama/pull/700.patch", "merged_at": "2023-10-06T14:15:42" }
rename llama.cpp `server.exe` to `ollama-runner`. This makes it easier to see that the subprocess is associated with ollama.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/700/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/700/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8647
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8647/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8647/comments
https://api.github.com/repos/ollama/ollama/issues/8647/events
https://github.com/ollama/ollama/issues/8647
2,817,099,829
I_kwDOJ0Z1Ps6n6Yg1
8,647
ollama[2075]: cuda driver library failed to get device context 999time=2025-01-29T03:21:43.859+01:00 level=WARN source=gpu.go:449 msg="error looking up nvidia GPU memory"
{ "login": "kurdo2000", "id": 137941417, "node_id": "U_kgDOCDjRqQ", "avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kurdo2000", "html_url": "https://github.com/kurdo2000", "followers_url": "https://api.github.com/users/kurdo2000/followers", "following_url": "https://api.github.com/users/kurdo2000/following{/other_user}", "gists_url": "https://api.github.com/users/kurdo2000/gists{/gist_id}", "starred_url": "https://api.github.com/users/kurdo2000/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kurdo2000/subscriptions", "organizations_url": "https://api.github.com/users/kurdo2000/orgs", "repos_url": "https://api.github.com/users/kurdo2000/repos", "events_url": "https://api.github.com/users/kurdo2000/events{/privacy}", "received_events_url": "https://api.github.com/users/kurdo2000/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
null
[]
null
2
2025-01-29T02:34:23
2025-01-30T03:09:02
2025-01-30T03:09:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi, is my GPU not supported or what's the error? +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.120 Driver Version: 550.120 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce GTX 1660 Off | 00000000:01:00.0 On | N/A | | 0% 48C P0 17W / 130W | 1602MiB / 6144MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | 0 N/A N/A 2627 G /usr/lib/xorg/Xorg 1032MiB | | 0 N/A N/A 2853 G /usr/bin/gnome-shell 79MiB | | 0 N/A N/A 3461 G /usr/libexec/xdg-desktop-portal-gnome 108MiB | | 0 N/A N/A 4054 G ...irefox/4793/usr/lib/firefox/firefox 203MiB | | 0 N/A N/A 25159 G /usr/bin/nautilus 40MiB | | 0 N/A N/A 106416 G ...erProcess --variations-seed-version 82MiB | +-----------------------------------------------------------------------------------------+ nvcc --version nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2024 NVIDIA Corporation Built on Tue_Oct_29_23:50:19_PDT_2024 Cuda compilation tools, release 12.6, V12.6.85 Build cuda_12.6.r12.6/compiler.35059454_0 If you still need more context, please ask. I'm new here. ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.7
{ "login": "kurdo2000", "id": 137941417, "node_id": "U_kgDOCDjRqQ", "avatar_url": "https://avatars.githubusercontent.com/u/137941417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kurdo2000", "html_url": "https://github.com/kurdo2000", "followers_url": "https://api.github.com/users/kurdo2000/followers", "following_url": "https://api.github.com/users/kurdo2000/following{/other_user}", "gists_url": "https://api.github.com/users/kurdo2000/gists{/gist_id}", "starred_url": "https://api.github.com/users/kurdo2000/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kurdo2000/subscriptions", "organizations_url": "https://api.github.com/users/kurdo2000/orgs", "repos_url": "https://api.github.com/users/kurdo2000/repos", "events_url": "https://api.github.com/users/kurdo2000/events{/privacy}", "received_events_url": "https://api.github.com/users/kurdo2000/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8647/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8647/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8067
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8067/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8067/comments
https://api.github.com/repos/ollama/ollama/issues/8067/events
https://github.com/ollama/ollama/issues/8067
2,735,177,482
I_kwDOJ0Z1Ps6jB38K
8,067
have a ques
{ "login": "17692019663", "id": 57002321, "node_id": "MDQ6VXNlcjU3MDAyMzIx", "avatar_url": "https://avatars.githubusercontent.com/u/57002321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/17692019663", "html_url": "https://github.com/17692019663", "followers_url": "https://api.github.com/users/17692019663/followers", "following_url": "https://api.github.com/users/17692019663/following{/other_user}", "gists_url": "https://api.github.com/users/17692019663/gists{/gist_id}", "starred_url": "https://api.github.com/users/17692019663/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/17692019663/subscriptions", "organizations_url": "https://api.github.com/users/17692019663/orgs", "repos_url": "https://api.github.com/users/17692019663/repos", "events_url": "https://api.github.com/users/17692019663/events{/privacy}", "received_events_url": "https://api.github.com/users/17692019663/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" }, { "id": 6677677816, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgVG-A", "url": "https://api.github.com/repos/ollama/ollama/labels/docker", "name": "docker", "color": "0052CC", "default": false, "description": "Issues relating to using ollama in containers" } ]
closed
false
null
[]
null
2
2024-12-12T08:17:34
2024-12-23T08:12:10
2024-12-23T08:12:10
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? C:\Users\Administrator> docker exec -it ollama ollama run llama3.2:1b pulling manifest pulling 74701a8c35f6... 100% ▕████████████████████████████████████████████████████████▏ 1.3 GB pulling 966de95ca8a6... 100% ▕████████████████████████████████████████████████████████▏ 1.4 KB pulling fcc5a6bec9da... 100% ▕████████████████████████████████████████████████████████▏ 7.7 KB pulling a70ff7e570d9... 100% ▕████████████████████████████████████████████████████████▏ 6.0 KB pulling 4f659a1e86d7... 100% ▕████████████████████████████████████████████████████████▏ 485 B verifying sha256 digest writing manifest success Error: error starting the external llama server: fork/exec /usr/lib/ollama/runners/cpu_avx2/ollama_llama_server: exec format error ### OS Windows ### GPU Intel ### CPU Intel ### Ollama version _No response_
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8067/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/134
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/134/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/134/comments
https://api.github.com/repos/ollama/ollama/issues/134/events
https://github.com/ollama/ollama/pull/134
1,813,267,553
PR_kwDOJ0Z1Ps5V9on7
134
Update images.go
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-07-20T06:34:19
2023-07-20T06:46:19
2023-07-20T06:46:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/134", "html_url": "https://github.com/ollama/ollama/pull/134", "diff_url": "https://github.com/ollama/ollama/pull/134.diff", "patch_url": "https://github.com/ollama/ollama/pull/134.patch", "merged_at": "2023-07-20T06:46:02" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/134/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/134/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3226
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3226/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3226/comments
https://api.github.com/repos/ollama/ollama/issues/3226/events
https://github.com/ollama/ollama/issues/3226
2,192,612,235
I_kwDOJ0Z1Ps6CsJuL
3,226
Cannot create Modelfile on Windows Shell
{ "login": "worras2019", "id": 52268635, "node_id": "MDQ6VXNlcjUyMjY4NjM1", "avatar_url": "https://avatars.githubusercontent.com/u/52268635?v=4", "gravatar_id": "", "url": "https://api.github.com/users/worras2019", "html_url": "https://github.com/worras2019", "followers_url": "https://api.github.com/users/worras2019/followers", "following_url": "https://api.github.com/users/worras2019/following{/other_user}", "gists_url": "https://api.github.com/users/worras2019/gists{/gist_id}", "starred_url": "https://api.github.com/users/worras2019/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/worras2019/subscriptions", "organizations_url": "https://api.github.com/users/worras2019/orgs", "repos_url": "https://api.github.com/users/worras2019/repos", "events_url": "https://api.github.com/users/worras2019/events{/privacy}", "received_events_url": "https://api.github.com/users/worras2019/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-03-18T15:57:03
2024-03-19T08:50:15
2024-03-19T08:50:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The following is not working in WIndows Shell. Windows Shell doesn't recognize FROM. -------------------------------------------------- FROM llama2 # set the temperature to 1 [higher is more creative, lower is more coherent] PARAMETER temperature 1 # set the system message SYSTEM """ You are Mario from Super Mario Bros. Answer as Mario, the assistant, only. """ ### What did you expect to see? Cannot create Modelfile ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS Windows ### Architecture _No response_ ### Platform _No response_ ### Ollama version _No response_ ### GPU _No response_ ### GPU info _No response_ ### CPU _No response_ ### Other software _No response_
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3226/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3226/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/2788
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2788/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2788/comments
https://api.github.com/repos/ollama/ollama/issues/2788/events
https://github.com/ollama/ollama/issues/2788
2,157,663,506
I_kwDOJ0Z1Ps6Am1US
2,788
Bug: LLaVA 1.6 34b not respecting initial user prompt
{ "login": "mobilemike", "id": 30590, "node_id": "MDQ6VXNlcjMwNTkw", "avatar_url": "https://avatars.githubusercontent.com/u/30590?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mobilemike", "html_url": "https://github.com/mobilemike", "followers_url": "https://api.github.com/users/mobilemike/followers", "following_url": "https://api.github.com/users/mobilemike/following{/other_user}", "gists_url": "https://api.github.com/users/mobilemike/gists{/gist_id}", "starred_url": "https://api.github.com/users/mobilemike/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mobilemike/subscriptions", "organizations_url": "https://api.github.com/users/mobilemike/orgs", "repos_url": "https://api.github.com/users/mobilemike/repos", "events_url": "https://api.github.com/users/mobilemike/events{/privacy}", "received_events_url": "https://api.github.com/users/mobilemike/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-27T21:47:53
2024-02-29T19:30:15
2024-02-29T19:30:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
M2 Max MBP 96GB RAM Ollama 0.1.27 Compared against llama.cpp CLI @b11a93d (same as Ollama version) Problem: When using the Ollama CLI or API with an image, the initial text prompt isn't respected. Examples like the one on the [model page](https://ollama.com/library/llava) typically show prompts like "What is in this picture?". However, when changing the initial prompt to something like "Is this image of a llama?" or "How many animals are in this picture?" or even "Ignore the image and tell me the meaning of life", the output is typically a description of the image. When using the llama.cpp CLI however, these prompts are followed as expected. When using chat completions in Ollama, a followup question _does_ work properly, so after initially being ignored you can get the expected output on a second attempt. My suspicion is that this behavior is largely unnoticed, as the default examples are asking for a description and one is being returned. However, this is masking the fact that text prompts used in conjunction with image prompts aren't being properly utilized. As an aside, LM Studio suffers from the same issue. The below examples use the same image as the one base64 encoded in the above model page CLI example. Ollama example: ``` ❯ ollama run llava:34b-v1.6-q6_K >>> /set parameter temperature 0.2 Set parameter 'temperature' to '0.2' >>> How many animals are in this picture? /Users/mike/Downloads/llama.png Added image '/Users/mike/Downloads/llama.png' The image you've provided appears to be a cartoon or illustration of an animal character. It looks like a cute, stylized depiction of a pig with a happy expression and waving its hand as if saying hello or goodbye. The art style is simplistic and playful, which is common in many modern cartoons and emojis. >>> How many animals are in this picture? There is only one animal in this picture, which is the cute pig character. ``` llama.cpp example: ``` ❯ ./llava-cli -m ../llm-models/cmp-nct/llava-1.6-gguf/ggml-yi-34b-f16-q_5_k.gguf --mmproj ../llm-models/cmp-nct/llava-1.6-gguf/mmproj-llava-34b-f16-q6_k.gguf --image ~/Downloads/llama.png --temp 0.2 -e -p '<|im_start|>system\n<|im_end|><|im_start|>user\n<image>\nHow many animals are in this picture?<|im_end|><|im_start|>assistant\n' clip_model_load: model name: vit-large336-custom clip_model_load: description: image encoder for LLaVA clip_model_load: GGUF version: 3 clip_model_load: alignment: 32 clip_model_load: n_tensors: 378 clip_model_load: n_kv: 26 clip_model_load: ftype: q6_K ... system_prompt: <|im_start|>system <|im_end|><|im_start|>user user_prompt: How many animals are in this picture?<|im_end|><|im_start|>assistant There is one animal in this picture, which appears to be a stylized drawing of a pig. ... ```
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2788/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2788/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4205
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4205/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4205/comments
https://api.github.com/repos/ollama/ollama/issues/4205/events
https://github.com/ollama/ollama/issues/4205
2,281,206,136
I_kwDOJ0Z1Ps6H-HF4
4,205
please support deepseek-ai/DeepSeek-V2-Chat
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/taozhiyuai/followers", "following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}", "gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}", "starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions", "organizations_url": "https://api.github.com/users/taozhiyuai/orgs", "repos_url": "https://api.github.com/users/taozhiyuai/repos", "events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}", "received_events_url": "https://api.github.com/users/taozhiyuai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
1
2024-05-06T15:54:12
2024-05-08T06:12:39
2024-05-08T06:12:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
please support deepseek-ai/DeepSeek-V2-Chat https://hf-mirror.com/deepseek-ai/DeepSeek-V2-Chat/tree/main
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/taozhiyuai/followers", "following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}", "gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}", "starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions", "organizations_url": "https://api.github.com/users/taozhiyuai/orgs", "repos_url": "https://api.github.com/users/taozhiyuai/repos", "events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}", "received_events_url": "https://api.github.com/users/taozhiyuai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4205/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4205/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2058
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2058/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2058/comments
https://api.github.com/repos/ollama/ollama/issues/2058/events
https://github.com/ollama/ollama/issues/2058
2,089,071,525
I_kwDOJ0Z1Ps58hLOl
2,058
Prompt Eval Count is 1 when image is included in multimodal request
{ "login": "Dillon-Yun", "id": 40805060, "node_id": "MDQ6VXNlcjQwODA1MDYw", "avatar_url": "https://avatars.githubusercontent.com/u/40805060?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Dillon-Yun", "html_url": "https://github.com/Dillon-Yun", "followers_url": "https://api.github.com/users/Dillon-Yun/followers", "following_url": "https://api.github.com/users/Dillon-Yun/following{/other_user}", "gists_url": "https://api.github.com/users/Dillon-Yun/gists{/gist_id}", "starred_url": "https://api.github.com/users/Dillon-Yun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Dillon-Yun/subscriptions", "organizations_url": "https://api.github.com/users/Dillon-Yun/orgs", "repos_url": "https://api.github.com/users/Dillon-Yun/repos", "events_url": "https://api.github.com/users/Dillon-Yun/events{/privacy}", "received_events_url": "https://api.github.com/users/Dillon-Yun/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-18T21:24:42
2024-10-23T17:46:39
2024-10-23T17:46:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When I hit the `/api/generate` endpoint for a multimodal model, such as bakllava, I am seeing that `prompt_eval_count == 1` when an image is included. Is this correct behavior? How can I retrieve the total amount of tokens in the request for a multimodal request?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2058/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2058/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5880
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5880/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5880/comments
https://api.github.com/repos/ollama/ollama/issues/5880/events
https://github.com/ollama/ollama/issues/5880
2,425,575,429
I_kwDOJ0Z1Ps6Qk1gF
5,880
Make Ollama available on Chocolatey
{ "login": "b-a0", "id": 25707742, "node_id": "MDQ6VXNlcjI1NzA3NzQy", "avatar_url": "https://avatars.githubusercontent.com/u/25707742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/b-a0", "html_url": "https://github.com/b-a0", "followers_url": "https://api.github.com/users/b-a0/followers", "following_url": "https://api.github.com/users/b-a0/following{/other_user}", "gists_url": "https://api.github.com/users/b-a0/gists{/gist_id}", "starred_url": "https://api.github.com/users/b-a0/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/b-a0/subscriptions", "organizations_url": "https://api.github.com/users/b-a0/orgs", "repos_url": "https://api.github.com/users/b-a0/repos", "events_url": "https://api.github.com/users/b-a0/events{/privacy}", "received_events_url": "https://api.github.com/users/b-a0/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
open
false
null
[]
null
2
2024-07-23T16:04:09
2024-09-05T19:52:37
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
The Windows installers works great, but requires that I manually download/update it from the ollama website. Would it be possible, probably when the Windows version has matured some more, to add ollama to [Chocolatey](https://community.chocolatey.org/)? That will probabably aid in making this an (even) easier to install package on Windows.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5880/reactions", "total_count": 4, "+1": 2, "-1": 2, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5880/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2377
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2377/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2377/comments
https://api.github.com/repos/ollama/ollama/issues/2377/events
https://github.com/ollama/ollama/pull/2377
2,121,583,534
PR_kwDOJ0Z1Ps5mL7sV
2,377
Bump llama.cpp to b2081
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-06T20:09:01
2024-02-07T20:04:41
2024-02-07T20:04:38
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2377", "html_url": "https://github.com/ollama/ollama/pull/2377", "diff_url": "https://github.com/ollama/ollama/pull/2377.diff", "patch_url": "https://github.com/ollama/ollama/pull/2377.patch", "merged_at": "2024-02-07T20:04:38" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2377/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2377/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3513
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3513/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3513/comments
https://api.github.com/repos/ollama/ollama/issues/3513/events
https://github.com/ollama/ollama/issues/3513
2,229,191,000
I_kwDOJ0Z1Ps6E3sFY
3,513
Stream: false is ignored in as of c863c6a96d01afda37b65f86eb14a1f04a3c7c47
{ "login": "eusebiu", "id": 414589, "node_id": "MDQ6VXNlcjQxNDU4OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eusebiu", "html_url": "https://github.com/eusebiu", "followers_url": "https://api.github.com/users/eusebiu/followers", "following_url": "https://api.github.com/users/eusebiu/following{/other_user}", "gists_url": "https://api.github.com/users/eusebiu/gists{/gist_id}", "starred_url": "https://api.github.com/users/eusebiu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eusebiu/subscriptions", "organizations_url": "https://api.github.com/users/eusebiu/orgs", "repos_url": "https://api.github.com/users/eusebiu/repos", "events_url": "https://api.github.com/users/eusebiu/events{/privacy}", "received_events_url": "https://api.github.com/users/eusebiu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-04-06T10:57:21
2024-04-13T07:22:27
2024-04-13T07:22:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Stream: false is ignored ### What did you expect to see? When stream is set to false, the response should contain the full result. ### Steps to reproduce Pull the latest code, compile, run (pull some model), run /api/generate with stream: false -> the result has no content (just some new lines) and "done: false". ### Are there any recent changes that introduced the issue? I think this https://github.com/ollama/ollama/commit/c863c6a96d01afda37b65f86eb14a1f04a3c7c47 breaks it (Tested before and after this commit). ### OS Windows ### Architecture amd64 ### Platform _No response_ ### Ollama version latest ### GPU Nvidia ### GPU info Quadro RTX 5000 with Max-Q Design ### CPU Intel ### Other software _No response_
{ "login": "eusebiu", "id": 414589, "node_id": "MDQ6VXNlcjQxNDU4OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/414589?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eusebiu", "html_url": "https://github.com/eusebiu", "followers_url": "https://api.github.com/users/eusebiu/followers", "following_url": "https://api.github.com/users/eusebiu/following{/other_user}", "gists_url": "https://api.github.com/users/eusebiu/gists{/gist_id}", "starred_url": "https://api.github.com/users/eusebiu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eusebiu/subscriptions", "organizations_url": "https://api.github.com/users/eusebiu/orgs", "repos_url": "https://api.github.com/users/eusebiu/repos", "events_url": "https://api.github.com/users/eusebiu/events{/privacy}", "received_events_url": "https://api.github.com/users/eusebiu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3513/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3513/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8117
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8117/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8117/comments
https://api.github.com/repos/ollama/ollama/issues/8117/events
https://github.com/ollama/ollama/pull/8117
2,742,219,356
PR_kwDOJ0Z1Ps6FWP5c
8,117
Examples(privategpt): update outdated langchain dependencies and add UnstructuredRSTLoader
{ "login": "KaymeKaydex", "id": 43151027, "node_id": "MDQ6VXNlcjQzMTUxMDI3", "avatar_url": "https://avatars.githubusercontent.com/u/43151027?v=4", "gravatar_id": "", "url": "https://api.github.com/users/KaymeKaydex", "html_url": "https://github.com/KaymeKaydex", "followers_url": "https://api.github.com/users/KaymeKaydex/followers", "following_url": "https://api.github.com/users/KaymeKaydex/following{/other_user}", "gists_url": "https://api.github.com/users/KaymeKaydex/gists{/gist_id}", "starred_url": "https://api.github.com/users/KaymeKaydex/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/KaymeKaydex/subscriptions", "organizations_url": "https://api.github.com/users/KaymeKaydex/orgs", "repos_url": "https://api.github.com/users/KaymeKaydex/repos", "events_url": "https://api.github.com/users/KaymeKaydex/events{/privacy}", "received_events_url": "https://api.github.com/users/KaymeKaydex/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-12-16T12:12:05
2025-01-13T19:26:24
2025-01-13T19:26:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8117", "html_url": "https://github.com/ollama/ollama/pull/8117", "diff_url": "https://github.com/ollama/ollama/pull/8117.diff", "patch_url": "https://github.com/ollama/ollama/pull/8117.patch", "merged_at": null }
This commit updates the langchain dependencies to the latest stable versions and adds the UnstructuredRSTLoader for loading RST files. - Update langchain dependencies to the latest stable versions. - Add UnstructuredRSTLoader for loading RST files.
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8117/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8117/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3074
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3074/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3074/comments
https://api.github.com/repos/ollama/ollama/issues/3074/events
https://github.com/ollama/ollama/issues/3074
2,181,146,552
I_kwDOJ0Z1Ps6CAae4
3,074
CUDA error: an illegal memory access was encountered
{ "login": "lizhichao999", "id": 34128722, "node_id": "MDQ6VXNlcjM0MTI4NzIy", "avatar_url": "https://avatars.githubusercontent.com/u/34128722?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lizhichao999", "html_url": "https://github.com/lizhichao999", "followers_url": "https://api.github.com/users/lizhichao999/followers", "following_url": "https://api.github.com/users/lizhichao999/following{/other_user}", "gists_url": "https://api.github.com/users/lizhichao999/gists{/gist_id}", "starred_url": "https://api.github.com/users/lizhichao999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lizhichao999/subscriptions", "organizations_url": "https://api.github.com/users/lizhichao999/orgs", "repos_url": "https://api.github.com/users/lizhichao999/repos", "events_url": "https://api.github.com/users/lizhichao999/events{/privacy}", "received_events_url": "https://api.github.com/users/lizhichao999/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
6
2024-03-12T09:35:45
2024-07-24T23:14:51
2024-07-24T23:14:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ollama run llama2 When I executed the command: ollama run llama2, an error occurred related to the data information. ![1710236043402](https://github.com/ollama/ollama/assets/34128722/098fb024-f6f6-419e-99a0-354d29eecd2d) I only have a 128G GPU. Do I need more than 2 GPUs? I can run it on my local laptop with 2 GPUs without any errors.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3074/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3074/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7704
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7704/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7704/comments
https://api.github.com/repos/ollama/ollama/issues/7704/events
https://github.com/ollama/ollama/issues/7704
2,665,123,889
I_kwDOJ0Z1Ps6e2pAx
7,704
On Windows 11 pro, it does work to right click "restart to update"
{ "login": "jack-liew", "id": 102201538, "node_id": "U_kgDOBhd4wg", "avatar_url": "https://avatars.githubusercontent.com/u/102201538?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jack-liew", "html_url": "https://github.com/jack-liew", "followers_url": "https://api.github.com/users/jack-liew/followers", "following_url": "https://api.github.com/users/jack-liew/following{/other_user}", "gists_url": "https://api.github.com/users/jack-liew/gists{/gist_id}", "starred_url": "https://api.github.com/users/jack-liew/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jack-liew/subscriptions", "organizations_url": "https://api.github.com/users/jack-liew/orgs", "repos_url": "https://api.github.com/users/jack-liew/repos", "events_url": "https://api.github.com/users/jack-liew/events{/privacy}", "received_events_url": "https://api.github.com/users/jack-liew/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
null
[]
null
1
2024-11-17T00:32:03
2024-11-18T22:39:53
2024-11-18T22:39:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? It does work to right click "restart to update". Thanks fo great work. ---------------------- ### OS Windows ### GPU Intel ### CPU Intel ### Ollama version 0.4.1
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7704/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7704/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8642
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8642/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8642/comments
https://api.github.com/repos/ollama/ollama/issues/8642/events
https://github.com/ollama/ollama/pull/8642
2,816,895,956
PR_kwDOJ0Z1Ps6JSssP
8,642
Update README with DeepSeek model in the examples
{ "login": "harsha-s", "id": 29120299, "node_id": "MDQ6VXNlcjI5MTIwMjk5", "avatar_url": "https://avatars.githubusercontent.com/u/29120299?v=4", "gravatar_id": "", "url": "https://api.github.com/users/harsha-s", "html_url": "https://github.com/harsha-s", "followers_url": "https://api.github.com/users/harsha-s/followers", "following_url": "https://api.github.com/users/harsha-s/following{/other_user}", "gists_url": "https://api.github.com/users/harsha-s/gists{/gist_id}", "starred_url": "https://api.github.com/users/harsha-s/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/harsha-s/subscriptions", "organizations_url": "https://api.github.com/users/harsha-s/orgs", "repos_url": "https://api.github.com/users/harsha-s/repos", "events_url": "https://api.github.com/users/harsha-s/events{/privacy}", "received_events_url": "https://api.github.com/users/harsha-s/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2025-01-28T23:10:31
2025-01-28T23:15:57
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8642", "html_url": "https://github.com/ollama/ollama/pull/8642", "diff_url": "https://github.com/ollama/ollama/pull/8642.diff", "patch_url": "https://github.com/ollama/ollama/pull/8642.patch", "merged_at": null }
Updated the README with the an example of DeepSeek model
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8642/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8642/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7306
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7306/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7306/comments
https://api.github.com/repos/ollama/ollama/issues/7306/events
https://github.com/ollama/ollama/pull/7306
2,603,930,359
PR_kwDOJ0Z1Ps5_Y0rC
7,306
integration: harden embedding test
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-22T00:15:14
2024-10-22T22:25:25
2024-10-22T22:25:22
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7306", "html_url": "https://github.com/ollama/ollama/pull/7306", "diff_url": "https://github.com/ollama/ollama/pull/7306.diff", "patch_url": "https://github.com/ollama/ollama/pull/7306.patch", "merged_at": "2024-10-22T22:25:22" }
Use cosine similarity to make the embeddings tests more robust
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7306/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7306/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1671
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1671/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1671/comments
https://api.github.com/repos/ollama/ollama/issues/1671/events
https://github.com/ollama/ollama/issues/1671
2,053,904,792
I_kwDOJ0Z1Ps56bBmY
1,671
Error: connect ECONNREFUSED 127.0.0.1:11434
{ "login": "LTtt456c", "id": 53040718, "node_id": "MDQ6VXNlcjUzMDQwNzE4", "avatar_url": "https://avatars.githubusercontent.com/u/53040718?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LTtt456c", "html_url": "https://github.com/LTtt456c", "followers_url": "https://api.github.com/users/LTtt456c/followers", "following_url": "https://api.github.com/users/LTtt456c/following{/other_user}", "gists_url": "https://api.github.com/users/LTtt456c/gists{/gist_id}", "starred_url": "https://api.github.com/users/LTtt456c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LTtt456c/subscriptions", "organizations_url": "https://api.github.com/users/LTtt456c/orgs", "repos_url": "https://api.github.com/users/LTtt456c/repos", "events_url": "https://api.github.com/users/LTtt456c/events{/privacy}", "received_events_url": "https://api.github.com/users/LTtt456c/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
8
2023-12-22T13:12:25
2024-06-14T10:53:14
2023-12-29T13:49:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello everyone! My ollama in My docker docker Start ollama command is docker run -e OLLAMA_HOST=0.0.0.0:11434 -d -v ollama serve -p 11434:11434 --name ollama ollama/ollama Then I in vscode open chatbot-ollama And then input npm run dev And then Report an error ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓ Here is the error log ↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓↓ PS G:\AI\chatbot-ollama> npm run dev chatbot-ollama@0.1.0 dev next dev ▲ Next.js 13.5.6 Local: http://localhost:3000/ ✓ Ready in 2.9s ○ Compiling / ... ✓ Compiled / in 3.3s (1652 modules) ⚠ Fast Refresh had to perform a full reload. Read more: https://nextjs.org/docs/messages/fast-refresh-reload ✓ Compiled in 1699ms (1652 modules) ✓ Compiled in 519ms (1652 modules) ✓ Compiled /api/models in 245ms (68 modules) [TypeError: fetch failed] { cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] { errno: -4078, code: 'ECONNREFUSED', syscall: 'connect', address: '127.0.0.1', port: 11434 } } ✓ Compiled in 620ms (1720 modules) [TypeError: fetch failed] { cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] { errno: -4078, code: 'ECONNREFUSED', syscall: 'connect', address: '127.0.0.1', port: 11434 } } [TypeError: fetch failed] { cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] { errno: -4078, code: 'ECONNREFUSED', syscall: 'connect', address: '127.0.0.1', port: 11434 } } [TypeError: fetch failed] { cause: [Error: connect ECONNREFUSED 127.0.0.1:11434] { errno: -4078, code: 'ECONNREFUSED', syscall: 'connect', address: '127.0.0.1', port: 11434 } }
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1671/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1671/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/756
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/756/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/756/comments
https://api.github.com/repos/ollama/ollama/issues/756/events
https://github.com/ollama/ollama/issues/756
1,938,044,665
I_kwDOJ0Z1Ps5zhDb5
756
Mistral - Failed To Load Model
{ "login": "mattdavenport", "id": 1127393, "node_id": "MDQ6VXNlcjExMjczOTM=", "avatar_url": "https://avatars.githubusercontent.com/u/1127393?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mattdavenport", "html_url": "https://github.com/mattdavenport", "followers_url": "https://api.github.com/users/mattdavenport/followers", "following_url": "https://api.github.com/users/mattdavenport/following{/other_user}", "gists_url": "https://api.github.com/users/mattdavenport/gists{/gist_id}", "starred_url": "https://api.github.com/users/mattdavenport/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mattdavenport/subscriptions", "organizations_url": "https://api.github.com/users/mattdavenport/orgs", "repos_url": "https://api.github.com/users/mattdavenport/repos", "events_url": "https://api.github.com/users/mattdavenport/events{/privacy}", "received_events_url": "https://api.github.com/users/mattdavenport/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-10-11T15:08:01
2023-10-11T15:40:14
2023-10-11T15:10:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm running macOS (Ventura 13.0.1) 16in. M1 2021. I am able to run all of the llama2 models just fine, but the following occurs when attempting to run the mistral model: ``` ~ % ollama pull mistral:latest pulling manifest pulling 6ae280299950... 100% |███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| (4.1/4.1 GB, 48 MB/s) pulling fede2d8d6c1f... 100% |████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| (29/29 B, 194 kB/s) pulling b96850d2e482... 100% |██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| (307/307 B, 1.4 MB/s) verifying sha256 digest writing manifest success ~ % ollama run mistral:latest >>> Hello Error: failed to load model ``` If this is still a WIP please close this issue. The only other information I could find is the following log entries: ``` 2023/10/11 10:25:47 images.go:1093: redirected to: https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/6a/6ae28029995007a3ee8d0b8556d50f3b59b831074cf19c84de87acf51fb54054/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20231011%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20231011T142546Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=386696851dae4763d830fc88c05381be653dab1e21243686e3180c01011644b6 2023/10/11 10:27:13 images.go:1061: success getting sha256:6ae28029995007a3ee8d0b8556d50f3b59b831074cf19c84de87acf51fb54054 2023/10/11 10:27:14 images.go:1093: redirected to: https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/fe/fede2d8d6c1f404b1db73b1cd26f7d5455ff2deeb737b5e2b339339dce2969d4/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20231011%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20231011T142714Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=4b1d208c4dcb6b20ae9727869c284c8ec0f77ee382ee975d96f50f1c358047e7 2023/10/11 10:27:14 images.go:1061: success getting sha256:fede2d8d6c1f404b1db73b1cd26f7d5455ff2deeb737b5e2b339339dce2969d4 2023/10/11 10:27:15 images.go:1093: redirected to: https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/b9/b96850d2e482b0d1af356eda4ac158af93e9b00e71363a9173d7b5480680bcf3/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20231011%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20231011T142715Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=a5cb6b9a865ca7264746ec67325b71ad80987c800bbeeadd43eef75a6e0363bc 2023/10/11 10:27:15 images.go:1061: success getting sha256:b96850d2e482b0d1af356eda4ac158af93e9b00e71363a9173d7b5480680bcf3 [GIN] 2023/10/11 - 10:27:18 | 200 | 1m33s | 127.0.0.1 | POST "/api/pull" llama.cpp: loading model from /Users/mattdavenport/.ollama/models/blobs/sha256:6ae28029995007a3ee8d0b8556d50f3b59b831074cf19c84de87acf51fb54054 error loading model: unknown (magic, version) combination: 46554747, 00000002; is this really a GGML file? llama_load_model_from_file: failed to load model [GIN] 2023/10/11 - 11:04:20 | 500 | 3.950083ms | 127.0.0.1 | POST "/api/generate" ``` Please let me know if I can provide any additional information here to help debug. Thanks!
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/756/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/756/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2363
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2363/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2363/comments
https://api.github.com/repos/ollama/ollama/issues/2363/events
https://github.com/ollama/ollama/pull/2363
2,119,594,984
PR_kwDOJ0Z1Ps5mFH5O
2,363
Move hub auth out to new package
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-05T22:39:49
2024-02-15T17:42:21
2024-02-15T17:42:17
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2363", "html_url": "https://github.com/ollama/ollama/pull/2363", "diff_url": "https://github.com/ollama/ollama/pull/2363.diff", "patch_url": "https://github.com/ollama/ollama/pull/2363.patch", "merged_at": null }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2363/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2363/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2514
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2514/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2514/comments
https://api.github.com/repos/ollama/ollama/issues/2514/events
https://github.com/ollama/ollama/issues/2514
2,136,760,146
I_kwDOJ0Z1Ps5_XF9S
2,514
Support for safetensors
{ "login": "gracehubai", "id": 159539863, "node_id": "U_kgDOCYJilw", "avatar_url": "https://avatars.githubusercontent.com/u/159539863?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gracehubai", "html_url": "https://github.com/gracehubai", "followers_url": "https://api.github.com/users/gracehubai/followers", "following_url": "https://api.github.com/users/gracehubai/following{/other_user}", "gists_url": "https://api.github.com/users/gracehubai/gists{/gist_id}", "starred_url": "https://api.github.com/users/gracehubai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gracehubai/subscriptions", "organizations_url": "https://api.github.com/users/gracehubai/orgs", "repos_url": "https://api.github.com/users/gracehubai/repos", "events_url": "https://api.github.com/users/gracehubai/events{/privacy}", "received_events_url": "https://api.github.com/users/gracehubai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-02-15T15:02:55
2024-02-21T05:38:43
2024-02-18T06:49:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Do we already support inferencing safetensors?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2514/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2514/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2214
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2214/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2214/comments
https://api.github.com/repos/ollama/ollama/issues/2214/events
https://github.com/ollama/ollama/pull/2214
2,102,794,214
PR_kwDOJ0Z1Ps5lMKh3
2,214
Detect lack of AVX and fallback to CPU mode
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-26T19:42:11
2024-01-26T20:06:47
2024-01-26T20:06:44
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2214", "html_url": "https://github.com/ollama/ollama/pull/2214", "diff_url": "https://github.com/ollama/ollama/pull/2214.diff", "patch_url": "https://github.com/ollama/ollama/pull/2214.patch", "merged_at": "2024-01-26T20:06:44" }
We build the GPU libraries with AVX enabled to ensure that if not all layers fit on the GPU we get better performance in a mixed mode. If the user is using a virtualization/emulation system that lacks AVX this used to result in an illegal instruction error and crash before this fix. Now we will report a warning in the server log, and just use CPU mode to ensure we don't crash. This should mitigate #2187.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2214/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2214/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8431
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8431/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8431/comments
https://api.github.com/repos/ollama/ollama/issues/8431/events
https://github.com/ollama/ollama/issues/8431
2,788,732,623
I_kwDOJ0Z1Ps6mOK7P
8,431
Embedding with bge-large:latest fails with "GGML_ASSERT(i01 >= 0 && i01 < ne01) failed"
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2025-01-15T03:07:10
2025-01-15T06:35:56
null
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
> @rick-github Thanks,I tried the "options":{"num_ctx":512} ,It works well! 💯 > but the logs always shows that > > ![Image](https://github.com/user-attachments/assets/ac0c0a66-aa0e-44df-bb25-fc387a09565f) _Originally posted by @Justin-12138 in [#7441](https://github.com/ollama/ollama/issues/7441#issuecomment-2591480334)_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8431/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8431/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1128
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1128/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1128/comments
https://api.github.com/repos/ollama/ollama/issues/1128/events
https://github.com/ollama/ollama/pull/1128
1,993,443,855
PR_kwDOJ0Z1Ps5fcrpD
1,128
FAQ: answer a few faq questions
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-11-14T19:43:51
2023-11-15T23:05:14
2023-11-15T23:05:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1128", "html_url": "https://github.com/ollama/ollama/pull/1128", "diff_url": "https://github.com/ollama/ollama/pull/1128.diff", "patch_url": "https://github.com/ollama/ollama/pull/1128.patch", "merged_at": "2023-11-15T23:05:13" }
null
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1128/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1128/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4247
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4247/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4247/comments
https://api.github.com/repos/ollama/ollama/issues/4247/events
https://github.com/ollama/ollama/pull/4247
2,284,495,403
PR_kwDOJ0Z1Ps5u0qMP
4,247
skip hidden files in list models handler
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-08T01:34:13
2024-05-08T02:01:46
2024-05-08T02:01:45
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4247", "html_url": "https://github.com/ollama/ollama/pull/4247", "diff_url": "https://github.com/ollama/ollama/pull/4247.diff", "patch_url": "https://github.com/ollama/ollama/pull/4247.patch", "merged_at": "2024-05-08T02:01:45" }
Hidden files on MacOS (ex: `.DS_Store`) cause the list command to file when not skipped: ``` ❯ ollama ls Error: unqualified name: ```
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4247/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4247/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4025
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4025/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4025/comments
https://api.github.com/repos/ollama/ollama/issues/4025/events
https://github.com/ollama/ollama/pull/4025
2,269,092,269
PR_kwDOJ0Z1Ps5uBD92
4,025
Update linux.md, suggest containerized install
{ "login": "ericcurtin", "id": 1694275, "node_id": "MDQ6VXNlcjE2OTQyNzU=", "avatar_url": "https://avatars.githubusercontent.com/u/1694275?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ericcurtin", "html_url": "https://github.com/ericcurtin", "followers_url": "https://api.github.com/users/ericcurtin/followers", "following_url": "https://api.github.com/users/ericcurtin/following{/other_user}", "gists_url": "https://api.github.com/users/ericcurtin/gists{/gist_id}", "starred_url": "https://api.github.com/users/ericcurtin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ericcurtin/subscriptions", "organizations_url": "https://api.github.com/users/ericcurtin/orgs", "repos_url": "https://api.github.com/users/ericcurtin/repos", "events_url": "https://api.github.com/users/ericcurtin/events{/privacy}", "received_events_url": "https://api.github.com/users/ericcurtin/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
2
2024-04-29T13:43:19
2024-05-06T22:54:56
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4025", "html_url": "https://github.com/ollama/ollama/pull/4025", "diff_url": "https://github.com/ollama/ollama/pull/4025.diff", "patch_url": "https://github.com/ollama/ollama/pull/4025.patch", "merged_at": null }
It's the simplest install process, once you run the one liner, everything just works.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4025/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4025/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5860
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5860/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5860/comments
https://api.github.com/repos/ollama/ollama/issues/5860/events
https://github.com/ollama/ollama/pull/5860
2,423,755,415
PR_kwDOJ0Z1Ps52I6Yu
5,860
auth: update auth
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-07-22T20:55:20
2024-09-04T13:43:34
2024-09-04T13:43:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5860", "html_url": "https://github.com/ollama/ollama/pull/5860", "diff_url": "https://github.com/ollama/ollama/pull/5860.diff", "patch_url": "https://github.com/ollama/ollama/pull/5860.patch", "merged_at": null }
pre cursor to setting up tests for https://github.com/ollama/ollama/pull/5441 moved initialize keypair to auth use initialize keypair implicitly when reading pubkey
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5860/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5860/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/511
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/511/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/511/comments
https://api.github.com/repos/ollama/ollama/issues/511/events
https://github.com/ollama/ollama/issues/511
1,891,545,545
I_kwDOJ0Z1Ps5wvrHJ
511
Support InternLM
{ "login": "vansinhu", "id": 60632596, "node_id": "MDQ6VXNlcjYwNjMyNTk2", "avatar_url": "https://avatars.githubusercontent.com/u/60632596?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vansinhu", "html_url": "https://github.com/vansinhu", "followers_url": "https://api.github.com/users/vansinhu/followers", "following_url": "https://api.github.com/users/vansinhu/following{/other_user}", "gists_url": "https://api.github.com/users/vansinhu/gists{/gist_id}", "starred_url": "https://api.github.com/users/vansinhu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vansinhu/subscriptions", "organizations_url": "https://api.github.com/users/vansinhu/orgs", "repos_url": "https://api.github.com/users/vansinhu/repos", "events_url": "https://api.github.com/users/vansinhu/events{/privacy}", "received_events_url": "https://api.github.com/users/vansinhu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
6
2023-09-12T02:09:05
2024-06-24T08:28:12
2024-03-11T19:32:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Dear ollama developer, Greetings! I am vansinhu, a community developer and volunteer at InternLM. [InternLM](https://github.com/InternLM/InternLM) is a large language model similar to llama2, and we look forward to InternLM being supported in ollama. If there are any challenges or inquiries regarding support for InternLM, please feel free to join our Discord discussion at https://discord.gg/gF9ezcmtM3. Best regards, vansinhu
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/511/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/511/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8383
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8383/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8383/comments
https://api.github.com/repos/ollama/ollama/issues/8383/events
https://github.com/ollama/ollama/pull/8383
2,781,700,019
PR_kwDOJ0Z1Ps6Hah72
8,383
fix: run container as non-root user
{ "login": "chaturvedi-kna", "id": 63336082, "node_id": "MDQ6VXNlcjYzMzM2MDgy", "avatar_url": "https://avatars.githubusercontent.com/u/63336082?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chaturvedi-kna", "html_url": "https://github.com/chaturvedi-kna", "followers_url": "https://api.github.com/users/chaturvedi-kna/followers", "following_url": "https://api.github.com/users/chaturvedi-kna/following{/other_user}", "gists_url": "https://api.github.com/users/chaturvedi-kna/gists{/gist_id}", "starred_url": "https://api.github.com/users/chaturvedi-kna/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chaturvedi-kna/subscriptions", "organizations_url": "https://api.github.com/users/chaturvedi-kna/orgs", "repos_url": "https://api.github.com/users/chaturvedi-kna/repos", "events_url": "https://api.github.com/users/chaturvedi-kna/events{/privacy}", "received_events_url": "https://api.github.com/users/chaturvedi-kna/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
1
2025-01-11T08:28:55
2025-01-19T02:54:33
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8383", "html_url": "https://github.com/ollama/ollama/pull/8383", "diff_url": "https://github.com/ollama/ollama/pull/8383.diff", "patch_url": "https://github.com/ollama/ollama/pull/8383.patch", "merged_at": null }
**Enable non-root user Container Deployment Support** **Description** This PR adds support for running ollama in rootless containers by properly setting up permissions for the `.ollama` directory. This change enables ollama to run in environments that enforce non-root execution such as: * OpenShift * Kubernetes with strict security contexts * Containers with user namespace remapping **Changes** * Added directory creation and permission setup in the final stage of Dockerfile * Set container to run as non-root user (UID 1001) * Configured group permissions to support random GID assignment **Testing Done** * Built image with podman * Deployed on OpenShift * Verified container starts successfully with non-root user * Confirmed `.ollama` directory is properly accessible **Previous Behavior** The container would fail to start in rootless environments with the error: ``` Couldn't find '/.ollama/id_ed25519'. Generating new private key. Error: could not create directory mkdir /.ollama: permission denied ``` **New Behavior** * Container starts successfully as non-root user * `.ollama` directory is created with proper permissions
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8383/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8383/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5306
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5306/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5306/comments
https://api.github.com/repos/ollama/ollama/issues/5306/events
https://github.com/ollama/ollama/pull/5306
2,375,887,192
PR_kwDOJ0Z1Ps5zqhZW
5,306
Do not reinstall the CLI tools if they are already installed on macOS
{ "login": "seanchristians", "id": 25487785, "node_id": "MDQ6VXNlcjI1NDg3Nzg1", "avatar_url": "https://avatars.githubusercontent.com/u/25487785?v=4", "gravatar_id": "", "url": "https://api.github.com/users/seanchristians", "html_url": "https://github.com/seanchristians", "followers_url": "https://api.github.com/users/seanchristians/followers", "following_url": "https://api.github.com/users/seanchristians/following{/other_user}", "gists_url": "https://api.github.com/users/seanchristians/gists{/gist_id}", "starred_url": "https://api.github.com/users/seanchristians/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/seanchristians/subscriptions", "organizations_url": "https://api.github.com/users/seanchristians/orgs", "repos_url": "https://api.github.com/users/seanchristians/repos", "events_url": "https://api.github.com/users/seanchristians/events{/privacy}", "received_events_url": "https://api.github.com/users/seanchristians/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-06-26T17:17:30
2024-09-06T17:49:40
2024-09-05T04:41:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5306", "html_url": "https://github.com/ollama/ollama/pull/5306", "diff_url": "https://github.com/ollama/ollama/pull/5306.diff", "patch_url": "https://github.com/ollama/ollama/pull/5306.patch", "merged_at": null }
Check `installed()` inside of the `install` function before running the cli setup commands. Fix #5305.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5306/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5306/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6165
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6165/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6165/comments
https://api.github.com/repos/ollama/ollama/issues/6165/events
https://github.com/ollama/ollama/issues/6165
2,447,319,803
I_kwDOJ0Z1Ps6R3yL7
6,165
No devices found using AMD gpus
{ "login": "cinglish", "id": 40436106, "node_id": "MDQ6VXNlcjQwNDM2MTA2", "avatar_url": "https://avatars.githubusercontent.com/u/40436106?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cinglish", "html_url": "https://github.com/cinglish", "followers_url": "https://api.github.com/users/cinglish/followers", "following_url": "https://api.github.com/users/cinglish/following{/other_user}", "gists_url": "https://api.github.com/users/cinglish/gists{/gist_id}", "starred_url": "https://api.github.com/users/cinglish/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cinglish/subscriptions", "organizations_url": "https://api.github.com/users/cinglish/orgs", "repos_url": "https://api.github.com/users/cinglish/repos", "events_url": "https://api.github.com/users/cinglish/events{/privacy}", "received_events_url": "https://api.github.com/users/cinglish/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg", "url": "https://api.github.com/repos/ollama/ollama/labels/linux", "name": "linux", "color": "516E70", "default": false, "description": "" }, { "id": 6433346500, "node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA", "url": "https://api.github.com/repos/ollama/ollama/labels/amd", "name": "amd", "color": "000000", "default": false, "description": "Issues relating to AMD GPUs and ROCm" }, { "id": 6677677816, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgVG-A", "url": "https://api.github.com/repos/ollama/ollama/labels/docker", "name": "docker", "color": "0052CC", "default": false, "description": "Issues relating to using ollama in containers" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-08-04T22:23:38
2024-09-05T19:03:53
2024-09-05T19:03:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Getting the following error when loading models with AMD gpus (Instinct MI60s): ``` rocBLAS error: Could not initialize Tensile host: No devices found ``` I have 4 devices allocated to the container and it seems to still discover them at startup: ``` time=2024-08-04T22:17:48.256Z level=INFO source=gpu.go:205 msg="looking for compatible GPUs" time=2024-08-04T22:17:48.264Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=0 gpu_type=gfx906 time=2024-08-04T22:17:48.264Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=1 time=2024-08-04T22:17:48.264Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=2 time=2024-08-04T22:17:48.264Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=3 time=2024-08-04T22:17:48.265Z level=WARN source=amd_linux.go:201 msg="amdgpu too old gfx000" gpu=4 time=2024-08-04T22:17:48.265Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=5 gpu_type=gfx906 time=2024-08-04T22:17:48.266Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=6 gpu_type=gfx906 time=2024-08-04T22:17:48.268Z level=INFO source=amd_linux.go:345 msg="amdgpu is supported" gpu=7 gpu_type=gfx906 ``` It is working fine with the `0.3.1-rocm` image, but seeing this behavior in the newest versions. ### OS Docker ### GPU AMD ### CPU Intel ### Ollama version 0.3.2
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6165/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6165/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2989
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2989/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2989/comments
https://api.github.com/repos/ollama/ollama/issues/2989/events
https://github.com/ollama/ollama/pull/2989
2,174,611,866
PR_kwDOJ0Z1Ps5pAaFv
2,989
CI updates
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-03-07T19:26:30
2024-03-09T21:20:27
2024-03-09T21:20:24
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2989", "html_url": "https://github.com/ollama/ollama/pull/2989", "diff_url": "https://github.com/ollama/ollama/pull/2989.diff", "patch_url": "https://github.com/ollama/ollama/pull/2989.patch", "merged_at": null }
Not ready for review yet...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2989/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2989/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/476
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/476/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/476/comments
https://api.github.com/repos/ollama/ollama/issues/476/events
https://github.com/ollama/ollama/pull/476
1,884,716,024
PR_kwDOJ0Z1Ps5ZuHR7
476
tighten up the error string for `ollama show` flags
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-06T20:37:08
2023-09-06T20:38:50
2023-09-06T20:38:49
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/476", "html_url": "https://github.com/ollama/ollama/pull/476", "diff_url": "https://github.com/ollama/ollama/pull/476.diff", "patch_url": "https://github.com/ollama/ollama/pull/476.patch", "merged_at": "2023-09-06T20:38:49" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/476/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/476/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3018
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3018/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3018/comments
https://api.github.com/repos/ollama/ollama/issues/3018/events
https://github.com/ollama/ollama/pull/3018
2,177,073,503
PR_kwDOJ0Z1Ps5pIy3-
3,018
Add allowed host middleware and remove `workDir` middleware
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-09T06:14:34
2024-03-09T06:23:48
2024-03-09T06:23:47
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3018", "html_url": "https://github.com/ollama/ollama/pull/3018", "diff_url": "https://github.com/ollama/ollama/pull/3018.diff", "patch_url": "https://github.com/ollama/ollama/pull/3018.patch", "merged_at": "2024-03-09T06:23:47" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3018/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3018/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/638
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/638/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/638/comments
https://api.github.com/repos/ollama/ollama/issues/638/events
https://github.com/ollama/ollama/pull/638
1,918,206,635
PR_kwDOJ0Z1Ps5beytn
638
add cuda to windows generation
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-28T20:15:47
2023-10-20T16:44:24
2023-09-29T16:53:14
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/638", "html_url": "https://github.com/ollama/ollama/pull/638", "diff_url": "https://github.com/ollama/ollama/pull/638.diff", "patch_url": "https://github.com/ollama/ollama/pull/638.patch", "merged_at": null }
Still a bunch of stuff to do for packaging this for distribution, but I tested building and running this locally and it works along with #637 so may as well get it in.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/638/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/638/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3616
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3616/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3616/comments
https://api.github.com/repos/ollama/ollama/issues/3616/events
https://github.com/ollama/ollama/issues/3616
2,240,686,081
I_kwDOJ0Z1Ps6FjigB
3,616
Richer grammars
{ "login": "tezlm", "id": 56274831, "node_id": "MDQ6VXNlcjU2Mjc0ODMx", "avatar_url": "https://avatars.githubusercontent.com/u/56274831?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tezlm", "html_url": "https://github.com/tezlm", "followers_url": "https://api.github.com/users/tezlm/followers", "following_url": "https://api.github.com/users/tezlm/following{/other_user}", "gists_url": "https://api.github.com/users/tezlm/gists{/gist_id}", "starred_url": "https://api.github.com/users/tezlm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tezlm/subscriptions", "organizations_url": "https://api.github.com/users/tezlm/orgs", "repos_url": "https://api.github.com/users/tezlm/repos", "events_url": "https://api.github.com/users/tezlm/events{/privacy}", "received_events_url": "https://api.github.com/users/tezlm/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
9
2024-04-12T18:23:12
2024-12-05T00:37:28
2024-12-05T00:37:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? Being able to specify grammars is great, but it seems a bit underutilized at the moment. This is mostly a thought dump on how it could be improved from experimentation... ### How should we solve this? - Using llama.cpp grammar directly would be pretty powerful and nice to have - Specifying jsonschema for json. Llama.cpp json is usually forced into a specific key order and ollama json isn't schema'd at all - Changing the format on the fly is useful, but I think it would be nice to have a way to specify a grammar in the Modelfile. ### What is the impact of not solving this? Not having either of the first two ideas is annoying, since there's no way to guarantee that a model generates a response in a format I want. The third idea allows one to make a "llm api", where a model generates a specific response every time (imagine bundling a "summary llm" that always responds with `{"summary":"..."}` as a Modelfile.) ### Anything else? _No response_
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3616/reactions", "total_count": 9, "+1": 9, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3616/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8624
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8624/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8624/comments
https://api.github.com/repos/ollama/ollama/issues/8624/events
https://github.com/ollama/ollama/issues/8624
2,814,669,000
I_kwDOJ0Z1Ps6nxHDI
8,624
Deepseek 80% size reduction
{ "login": "gileneusz", "id": 34601970, "node_id": "MDQ6VXNlcjM0NjAxOTcw", "avatar_url": "https://avatars.githubusercontent.com/u/34601970?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gileneusz", "html_url": "https://github.com/gileneusz", "followers_url": "https://api.github.com/users/gileneusz/followers", "following_url": "https://api.github.com/users/gileneusz/following{/other_user}", "gists_url": "https://api.github.com/users/gileneusz/gists{/gist_id}", "starred_url": "https://api.github.com/users/gileneusz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gileneusz/subscriptions", "organizations_url": "https://api.github.com/users/gileneusz/orgs", "repos_url": "https://api.github.com/users/gileneusz/repos", "events_url": "https://api.github.com/users/gileneusz/events{/privacy}", "received_events_url": "https://api.github.com/users/gileneusz/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2025-01-28T05:27:54
2025-01-28T22:28:24
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
New quants done by unsloth.ai: | MoE Bits | Disk Size | Type | Quality | Link | Down_proj | |-----------|-----------|----------|---------|------------------------------------------------------------------------------------------------------------------------------------------------|----------------| | 1.58-bit | 131GB | IQ1_S | Fair | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-IQ1_S) | 2.06/1.56bit | | 1.73-bit | 158GB | IQ1_M | Good | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-IQ1_M) | 2.06bit | | 2.22-bit | 183GB | IQ2_XXS | Better | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-IQ2_XXS) | 2.5/2.06bit | | 2.51-bit | 212GB | Q2_K_XL | Best | [Link](https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-Q2_K_XL) | 3.5/2.5bit | please consider adding them https://unsloth.ai/blog/deepseekr1-dynamic thanks!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8624/reactions", "total_count": 8, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 8, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8624/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/59
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/59/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/59/comments
https://api.github.com/repos/ollama/ollama/issues/59/events
https://github.com/ollama/ollama/issues/59
1,795,166,966
I_kwDOJ0Z1Ps5rABL2
59
generate pauses after about 50 tokens
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2023-07-08T23:42:53
2023-07-13T02:15:17
2023-07-13T02:15:17
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Generation will get paused after about 50 tokens being provided ``` % ollama run orca >>> Write a review of the restaurant "five guys" As an AI assistant, I cannot write a biased or subjective review, but I can provide you with some general information about the restaurant "Five Guys". Five Guys is an American fast-food chain that primarily serves hamburgers, fries, <pause here> ```
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/59/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/59/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5278
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5278/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5278/comments
https://api.github.com/repos/ollama/ollama/issues/5278/events
https://github.com/ollama/ollama/issues/5278
2,373,277,484
I_kwDOJ0Z1Ps6NdVcs
5,278
Is it possible to start llama server through dynamic dependency library?
{ "login": "leeyiding", "id": 49098278, "node_id": "MDQ6VXNlcjQ5MDk4Mjc4", "avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leeyiding", "html_url": "https://github.com/leeyiding", "followers_url": "https://api.github.com/users/leeyiding/followers", "following_url": "https://api.github.com/users/leeyiding/following{/other_user}", "gists_url": "https://api.github.com/users/leeyiding/gists{/gist_id}", "starred_url": "https://api.github.com/users/leeyiding/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leeyiding/subscriptions", "organizations_url": "https://api.github.com/users/leeyiding/orgs", "repos_url": "https://api.github.com/users/leeyiding/repos", "events_url": "https://api.github.com/users/leeyiding/events{/privacy}", "received_events_url": "https://api.github.com/users/leeyiding/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-06-25T17:43:21
2024-07-03T02:36:33
2024-07-03T02:36:33
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I'm trying to run Ollama in [Nanos Unikernel](https://github.com/nanovms/nanos), Unikernel a single-process operating system. I found that in https://github.com/ollama/ollama/commit/58d95cc9bd446a8209e7388a96c70367cbafd653, the running mode of llama server changed from loading dynamic dependency libraries to running through subprocess. Therefore, before version 0.1.32, I could run Ollama normally in Nanos, but it could not run normally in later versions. I would like to ask whether it is possible to continue to provide the method of loading llama server as a dynamic dependency library and avoid using subprocess. Perhaps the two methods can coexist and users can choose by themselves. Looking forward to your reply, thank you very much.
{ "login": "leeyiding", "id": 49098278, "node_id": "MDQ6VXNlcjQ5MDk4Mjc4", "avatar_url": "https://avatars.githubusercontent.com/u/49098278?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leeyiding", "html_url": "https://github.com/leeyiding", "followers_url": "https://api.github.com/users/leeyiding/followers", "following_url": "https://api.github.com/users/leeyiding/following{/other_user}", "gists_url": "https://api.github.com/users/leeyiding/gists{/gist_id}", "starred_url": "https://api.github.com/users/leeyiding/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leeyiding/subscriptions", "organizations_url": "https://api.github.com/users/leeyiding/orgs", "repos_url": "https://api.github.com/users/leeyiding/repos", "events_url": "https://api.github.com/users/leeyiding/events{/privacy}", "received_events_url": "https://api.github.com/users/leeyiding/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5278/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5278/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1579
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1579/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1579/comments
https://api.github.com/repos/ollama/ollama/issues/1579/events
https://github.com/ollama/ollama/issues/1579
2,046,156,422
I_kwDOJ0Z1Ps559d6G
1,579
Error while running ollama locally.
{ "login": "nehalmathew1996", "id": 82964954, "node_id": "MDQ6VXNlcjgyOTY0OTU0", "avatar_url": "https://avatars.githubusercontent.com/u/82964954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nehalmathew1996", "html_url": "https://github.com/nehalmathew1996", "followers_url": "https://api.github.com/users/nehalmathew1996/followers", "following_url": "https://api.github.com/users/nehalmathew1996/following{/other_user}", "gists_url": "https://api.github.com/users/nehalmathew1996/gists{/gist_id}", "starred_url": "https://api.github.com/users/nehalmathew1996/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nehalmathew1996/subscriptions", "organizations_url": "https://api.github.com/users/nehalmathew1996/orgs", "repos_url": "https://api.github.com/users/nehalmathew1996/repos", "events_url": "https://api.github.com/users/nehalmathew1996/events{/privacy}", "received_events_url": "https://api.github.com/users/nehalmathew1996/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
11
2023-12-18T09:28:17
2024-04-05T21:14:06
2024-02-01T23:19:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ConnectionError: HTTPConnectionPool(host='localhost', port=11434): Max retries exceeded with url: /api/generate/ (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x000002298AE1EF50>: Failed to establish a new connection: [WinError 10061] No connection could be made because the target machine actively refused it'))
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1579/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1579/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7394
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7394/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7394/comments
https://api.github.com/repos/ollama/ollama/issues/7394/events
https://github.com/ollama/ollama/issues/7394
2,617,770,363
I_kwDOJ0Z1Ps6cCAF7
7,394
The Open WebUI generate unsense text, but the cli terminal can chat normally with the same hosted model
{ "login": "QiuJYWX", "id": 68810978, "node_id": "MDQ6VXNlcjY4ODEwOTc4", "avatar_url": "https://avatars.githubusercontent.com/u/68810978?v=4", "gravatar_id": "", "url": "https://api.github.com/users/QiuJYWX", "html_url": "https://github.com/QiuJYWX", "followers_url": "https://api.github.com/users/QiuJYWX/followers", "following_url": "https://api.github.com/users/QiuJYWX/following{/other_user}", "gists_url": "https://api.github.com/users/QiuJYWX/gists{/gist_id}", "starred_url": "https://api.github.com/users/QiuJYWX/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/QiuJYWX/subscriptions", "organizations_url": "https://api.github.com/users/QiuJYWX/orgs", "repos_url": "https://api.github.com/users/QiuJYWX/repos", "events_url": "https://api.github.com/users/QiuJYWX/events{/privacy}", "received_events_url": "https://api.github.com/users/QiuJYWX/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
9
2024-10-28T09:06:43
2024-11-13T22:12:46
2024-11-13T22:12:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The results generated by cli terminal: ![cli](https://github.com/user-attachments/assets/aa1ea3c5-bcd0-45e0-b9e4-480aa81036ec) The results generated by open webui: ![webui](https://github.com/user-attachments/assets/81f236c9-517c-4fc5-a3e3-6c9f284c7555) ![webui2](https://github.com/user-attachments/assets/4f35c1b5-a422-402c-82bc-eecdeba35934) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.14
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7394/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7394/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8316
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8316/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8316/comments
https://api.github.com/repos/ollama/ollama/issues/8316/events
https://github.com/ollama/ollama/issues/8316
2,770,077,246
I_kwDOJ0Z1Ps6lHAY-
8,316
Feature Request: Limit Ollama to use only GPU RAM (VRAM), not system RAM
{ "login": "mlibre", "id": 8473036, "node_id": "MDQ6VXNlcjg0NzMwMzY=", "avatar_url": "https://avatars.githubusercontent.com/u/8473036?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mlibre", "html_url": "https://github.com/mlibre", "followers_url": "https://api.github.com/users/mlibre/followers", "following_url": "https://api.github.com/users/mlibre/following{/other_user}", "gists_url": "https://api.github.com/users/mlibre/gists{/gist_id}", "starred_url": "https://api.github.com/users/mlibre/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mlibre/subscriptions", "organizations_url": "https://api.github.com/users/mlibre/orgs", "repos_url": "https://api.github.com/users/mlibre/repos", "events_url": "https://api.github.com/users/mlibre/events{/privacy}", "received_events_url": "https://api.github.com/users/mlibre/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
4
2025-01-06T08:02:41
2025-01-24T09:45:19
2025-01-24T09:45:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### Feature Request **Description**: I would like to request a feature that allows Ollama to be configured to use only GPU RAM, without utilizing CPU or system RAM. This would be beneficial for systems with limited RAM resources and high GPU RAM (VRAM) memory availability. **Benefits**: - Optimized performance on GPU-heavy systems - Reduced CPU load and system RAM usage - Enhanced efficiency for specific workloads **Example Use Case**: In scenarios where the system has multiple GPUs with high memory capacity but limited CPU and system RAM, this feature would enable Ollama to leverage the full potential of the GPUs without being constrained by CPU or system RAM limitations. **Environment Variables/Configuration**: - An environment variable or configuration setting to enforce GPU-only memory usage. - Example: `OLLAMA_USE_GPU_ONLY=1` **Related Files and Code**: - Documentation: [docs/gpu.md](https://github.com/ollama/ollama/blob/main/docs/gpu.md)
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8316/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8316/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8562
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8562/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8562/comments
https://api.github.com/repos/ollama/ollama/issues/8562/events
https://github.com/ollama/ollama/issues/8562
2,809,089,096
I_kwDOJ0Z1Ps6nb0xI
8,562
None value being returned from with_structured_output request
{ "login": "jonmach", "id": 7843429, "node_id": "MDQ6VXNlcjc4NDM0Mjk=", "avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonmach", "html_url": "https://github.com/jonmach", "followers_url": "https://api.github.com/users/jonmach/followers", "following_url": "https://api.github.com/users/jonmach/following{/other_user}", "gists_url": "https://api.github.com/users/jonmach/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonmach/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonmach/subscriptions", "organizations_url": "https://api.github.com/users/jonmach/orgs", "repos_url": "https://api.github.com/users/jonmach/repos", "events_url": "https://api.github.com/users/jonmach/events{/privacy}", "received_events_url": "https://api.github.com/users/jonmach/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
6
2025-01-24T10:26:05
2025-01-27T10:15:42
2025-01-24T13:27:20
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Python versions are: langchain 0.3.15 langchain-community 0.3.15 langchain-core 0.3.31 langchain-ollama 0.2.2 ollama 0.4.7 Running ollama 0.5.7 (pip install -U ollama did not increase the version beyond 0.4.7) --- Using **with_structured_output()** seems to work for a very simple example such as the following: ``` from langchain_ollama import ChatOllama from typing import Optional from pydantic import BaseModel, Field class Person(BaseModel): name: str age: int llm = ChatOllama( model="qwen2.5:1.5b", temperature=0, ).with_structured_output(Person) llm.invoke("Erick 27") ``` However, for a more complex requirement, it fails with Ollama returning a value of None. ``` from pydantic import BaseModel, Field from typing import Optional from openai import OpenAI from langchain_ollama import ChatOllama # Define the output model class Experience(BaseModel): company: str = Field(..., description="The name of the company.") position: str = Field(..., description="The job title held at the company.") start_date: str = Field(..., description="The date when you started working at the company.") end_date: str = Field(..., description="The date when you left the company. If still employed, use 'Present'.") class Education(BaseModel): institution_name: str = Field(..., description="The name of the educational institution.") degree: str = Field(..., description="The degree obtained from the institution.") start_date: str = Field(..., description="The date when you started attending school at the institution.") end_date: str = Field(..., description="The date when you graduated. If still enrolled, use 'Present'.") class Resume(BaseModel): full_name: str = Field(..., description="The full name of the person on the resume.") contact_email: str = Field(..., description="The email address for contacting the person.") phone_number: str = Field(..., description="The phone number for contacting the person.") summary: str = Field(..., description="A brief summary of the person's career highlights.") experience: Optional[list[Experience]] = Field([], description="List of experiences held by the person.") education: Optional[list[Education]] = Field([], description="List of educational institutions attended by the person.") with open('CVs/resume.md', 'r') as file: resume_data = file.read() verbose=True model = "qwen2.5:14b" prompt = f""" Analyse the following resume from the content between the triple backticks below: For the resume below, identify the following information: 1) Their personal details, including name, email, phone number and anything else they provide. 2) An overall summary of their experience to provide a general background. 3) A list of the companies they have worked for. This should include the company name, the dates they started and and ended working for the company, and the tasks and activities they carried out. 4) A list of universities or colleges that the person went to. This should include the name of the college the title of the qualification, and the dates they started and ended. The raw data is here: ```{resume_data}``` """ # print(prompt) llm = ChatOllama( model=model, num_ctx = 32000, timeout = 600, temperature=0.0, verboseness = verbose, response = "json") structured_llm = llm.with_structured_output(Resume) print("Calling LLM") response = structured_llm.invoke(prompt) print(response) ``` It also fails without the 'response = "json" included. I just get a None response. Oddly, this is not consistent. Sometimes, I get back a response, but it fails satisfying the Resume type. because it won't find education items. Even though Education is an optional type in the Resume class. For some reason, I cannot upload the small resume file, so here it is in cleartext: --- ``` ## **Professional Experience** ### **Senior Software Engineer** **Tech Innovators Inc.** _June 2015 – Present_ - Designed and implemented scalable microservices architecture for a SaaS platform, improving performance by 30%. - Led a team of 12 engineers, mentoring junior developers and conducting regular code reviews. - Integrated AI/ML capabilities into legacy systems, increasing operational efficiency by 20%. - Championed DevOps practices, reducing deployment times from days to hours. ### **Software Architect** **NextGen Solutions** _March 2010 – May 2015_ - Architected and delivered a real-time analytics platform for financial services, handling millions of transactions daily. - Migrated a monolithic system to a distributed microservices-based architecture, enabling faster feature delivery. - Partnered with product managers to define technical requirements and roadmap, aligning business goals with engineering efforts. ### **Lead Developer** **Alpha Development Corp.** _January 2005 – February 2010_ - Built a high-availability e-commerce platform that handled over 500,000 daily users. - Created APIs to integrate third-party payment gateways, enhancing user experience and reducing downtime. - Conducted performance optimizations that improved application speed by 40%. ### **Software Engineer** **CodeSphere LLC** _June 2000 – December 2004_ - Developed enterprise-grade web applications using Java and C++. - Automated internal processes, saving the company 15% in operational costs annually. - Collaborated with cross-functional teams to deliver projects on time and within budget. --- ## **Education** ### **Master of Science in Computer Science** **Massachusetts Institute of Technology** _August 1998 – May 2000_ ### **Bachelor of Science in Computer Science** **University of California, Berkeley** _August 1994 – May 1998_ --- ## **Skills** - Programming Languages: Python, Java, C++, JavaScript - Cloud Platforms: AWS, Azure, Google Cloud - Architecture: Microservices, Distributed Systems, RESTful APIs - Tools: Docker, Kubernetes, Terraform - Agile Development, DevOps, AI/ML Integration --- ## **Certifications** - AWS Certified Solutions Architect – Professional - Certified Kubernetes Administrator (CKA) - Certified ScrumMaster (CSM) --- ## **Contact** Feel free to reach out via email or phone for opportunities or collaboration. ``` ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.5.7
{ "login": "jonmach", "id": 7843429, "node_id": "MDQ6VXNlcjc4NDM0Mjk=", "avatar_url": "https://avatars.githubusercontent.com/u/7843429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonmach", "html_url": "https://github.com/jonmach", "followers_url": "https://api.github.com/users/jonmach/followers", "following_url": "https://api.github.com/users/jonmach/following{/other_user}", "gists_url": "https://api.github.com/users/jonmach/gists{/gist_id}", "starred_url": "https://api.github.com/users/jonmach/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jonmach/subscriptions", "organizations_url": "https://api.github.com/users/jonmach/orgs", "repos_url": "https://api.github.com/users/jonmach/repos", "events_url": "https://api.github.com/users/jonmach/events{/privacy}", "received_events_url": "https://api.github.com/users/jonmach/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8562/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8562/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/130
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/130/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/130/comments
https://api.github.com/repos/ollama/ollama/issues/130/events
https://github.com/ollama/ollama/pull/130
1,812,847,886
PR_kwDOJ0Z1Ps5V8NI1
130
vendor in progress bar and change to bytes instead of bibytes
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-07-19T22:54:28
2023-07-20T00:24:03
2023-07-20T00:24:03
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/130", "html_url": "https://github.com/ollama/ollama/pull/130", "diff_url": "https://github.com/ollama/ollama/pull/130.diff", "patch_url": "https://github.com/ollama/ollama/pull/130.patch", "merged_at": "2023-07-20T00:24:03" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/130/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/130/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4611
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4611/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4611/comments
https://api.github.com/repos/ollama/ollama/issues/4611/events
https://github.com/ollama/ollama/issues/4611
2,314,861,010
I_kwDOJ0Z1Ps6J-fnS
4,611
User comments on personal model page
{ "login": "razvanab", "id": 2854730, "node_id": "MDQ6VXNlcjI4NTQ3MzA=", "avatar_url": "https://avatars.githubusercontent.com/u/2854730?v=4", "gravatar_id": "", "url": "https://api.github.com/users/razvanab", "html_url": "https://github.com/razvanab", "followers_url": "https://api.github.com/users/razvanab/followers", "following_url": "https://api.github.com/users/razvanab/following{/other_user}", "gists_url": "https://api.github.com/users/razvanab/gists{/gist_id}", "starred_url": "https://api.github.com/users/razvanab/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/razvanab/subscriptions", "organizations_url": "https://api.github.com/users/razvanab/orgs", "repos_url": "https://api.github.com/users/razvanab/repos", "events_url": "https://api.github.com/users/razvanab/events{/privacy}", "received_events_url": "https://api.github.com/users/razvanab/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
open
false
null
[]
null
0
2024-05-24T09:05:07
2024-07-08T17:22:09
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I would like to suggest a section for user comments on the personal Ollama model page. This will help with better tuning the parameters for your version of the model.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4611/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4611/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4287
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4287/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4287/comments
https://api.github.com/repos/ollama/ollama/issues/4287/events
https://github.com/ollama/ollama/pull/4287
2,287,796,918
PR_kwDOJ0Z1Ps5u_yAg
4,287
suppress ollama_llama_server.exe blank command window popup
{ "login": "ashokgelal", "id": 401055, "node_id": "MDQ6VXNlcjQwMTA1NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/401055?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ashokgelal", "html_url": "https://github.com/ashokgelal", "followers_url": "https://api.github.com/users/ashokgelal/followers", "following_url": "https://api.github.com/users/ashokgelal/following{/other_user}", "gists_url": "https://api.github.com/users/ashokgelal/gists{/gist_id}", "starred_url": "https://api.github.com/users/ashokgelal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ashokgelal/subscriptions", "organizations_url": "https://api.github.com/users/ashokgelal/orgs", "repos_url": "https://api.github.com/users/ashokgelal/repos", "events_url": "https://api.github.com/users/ashokgelal/events{/privacy}", "received_events_url": "https://api.github.com/users/ashokgelal/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
10
2024-05-09T14:19:20
2025-01-29T16:31:20
2024-11-23T21:29:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4287", "html_url": "https://github.com/ollama/ollama/pull/4287", "diff_url": "https://github.com/ollama/ollama/pull/4287.diff", "patch_url": "https://github.com/ollama/ollama/pull/4287.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4287/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4287/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1827
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1827/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1827/comments
https://api.github.com/repos/ollama/ollama/issues/1827/events
https://github.com/ollama/ollama/issues/1827
2,068,731,697
I_kwDOJ0Z1Ps57Tlcx
1,827
Massive slowdown on v 0.1.18 vs 0.1.17 with same model on Intel Mac
{ "login": "pjv", "id": 327716, "node_id": "MDQ6VXNlcjMyNzcxNg==", "avatar_url": "https://avatars.githubusercontent.com/u/327716?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pjv", "html_url": "https://github.com/pjv", "followers_url": "https://api.github.com/users/pjv/followers", "following_url": "https://api.github.com/users/pjv/following{/other_user}", "gists_url": "https://api.github.com/users/pjv/gists{/gist_id}", "starred_url": "https://api.github.com/users/pjv/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pjv/subscriptions", "organizations_url": "https://api.github.com/users/pjv/orgs", "repos_url": "https://api.github.com/users/pjv/repos", "events_url": "https://api.github.com/users/pjv/events{/privacy}", "received_events_url": "https://api.github.com/users/pjv/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
8
2024-01-06T17:18:27
2024-04-15T01:15:54
2024-01-07T05:46:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
<img width="255" alt="SCR-20240106-kfri" src="https://github.com/jmorganca/ollama/assets/327716/c1b5ce5b-acd5-4c81-a59a-6db39ff6a257"> I don’t have exact timings but the same model (in this case, `deepseek-coder:6.7b-instruct-q4_K_S`) generates tokens roughly 5 times faster on 0.1.17 than on 0.1.18 on my Intel Mac. I upgraded to 0.1.18 and noticed the slowdown in token generation and then downgraded back to 0.1.17 and immediately saw the faster throughput I am accustomed to.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1827/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1827/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8108
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8108/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8108/comments
https://api.github.com/repos/ollama/ollama/issues/8108/events
https://github.com/ollama/ollama/issues/8108
2,740,794,394
I_kwDOJ0Z1Ps6jXTQa
8,108
Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64
{ "login": "dernikolas", "id": 150927066, "node_id": "U_kgDOCP722g", "avatar_url": "https://avatars.githubusercontent.com/u/150927066?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dernikolas", "html_url": "https://github.com/dernikolas", "followers_url": "https://api.github.com/users/dernikolas/followers", "following_url": "https://api.github.com/users/dernikolas/following{/other_user}", "gists_url": "https://api.github.com/users/dernikolas/gists{/gist_id}", "starred_url": "https://api.github.com/users/dernikolas/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dernikolas/subscriptions", "organizations_url": "https://api.github.com/users/dernikolas/orgs", "repos_url": "https://api.github.com/users/dernikolas/repos", "events_url": "https://api.github.com/users/dernikolas/events{/privacy}", "received_events_url": "https://api.github.com/users/dernikolas/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg", "url": "https://api.github.com/repos/ollama/ollama/labels/linux", "name": "linux", "color": "516E70", "default": false, "description": "" }, { "id": 6433346500, "node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA", "url": "https://api.github.com/repos/ollama/ollama/labels/amd", "name": "amd", "color": "000000", "default": false, "description": "Issues relating to AMD GPUs and ROCm" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-12-15T17:57:17
2025-01-06T17:54:55
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When I start llama3.2 with `ollama run llama3.2` this error spawns: `Error: llama runner process has terminated: error:/opt/rocm/lib/libhipblas.so.2: undefined symbol: rocblas_sgbmv_64` I am on Cachy OS with ollama and ollama-rocm installed from extra repo and a AMD RX6800 is the GPU ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.5.2
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8108/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8108/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6571
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6571/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6571/comments
https://api.github.com/repos/ollama/ollama/issues/6571/events
https://github.com/ollama/ollama/issues/6571
2,497,977,180
I_kwDOJ0Z1Ps6U5Btc
6,571
Impossible to connect to ollama locally from another pc
{ "login": "Wilnox23", "id": 82610998, "node_id": "MDQ6VXNlcjgyNjEwOTk4", "avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Wilnox23", "html_url": "https://github.com/Wilnox23", "followers_url": "https://api.github.com/users/Wilnox23/followers", "following_url": "https://api.github.com/users/Wilnox23/following{/other_user}", "gists_url": "https://api.github.com/users/Wilnox23/gists{/gist_id}", "starred_url": "https://api.github.com/users/Wilnox23/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wilnox23/subscriptions", "organizations_url": "https://api.github.com/users/Wilnox23/orgs", "repos_url": "https://api.github.com/users/Wilnox23/repos", "events_url": "https://api.github.com/users/Wilnox23/events{/privacy}", "received_events_url": "https://api.github.com/users/Wilnox23/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-08-30T18:45:36
2024-08-30T21:12:58
2024-08-30T21:11:07
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hello, I'm using ollama on windows but I can't connect to the server from another PC on the local network. I have already changed the following system environment variables. `OLLAMA_HOST=0.0.0.0` `OLLAMA_ORIGINS=*` Thank you for your assistance. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.8
{ "login": "Wilnox23", "id": 82610998, "node_id": "MDQ6VXNlcjgyNjEwOTk4", "avatar_url": "https://avatars.githubusercontent.com/u/82610998?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Wilnox23", "html_url": "https://github.com/Wilnox23", "followers_url": "https://api.github.com/users/Wilnox23/followers", "following_url": "https://api.github.com/users/Wilnox23/following{/other_user}", "gists_url": "https://api.github.com/users/Wilnox23/gists{/gist_id}", "starred_url": "https://api.github.com/users/Wilnox23/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Wilnox23/subscriptions", "organizations_url": "https://api.github.com/users/Wilnox23/orgs", "repos_url": "https://api.github.com/users/Wilnox23/repos", "events_url": "https://api.github.com/users/Wilnox23/events{/privacy}", "received_events_url": "https://api.github.com/users/Wilnox23/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6571/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6571/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2693
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2693/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2693/comments
https://api.github.com/repos/ollama/ollama/issues/2693/events
https://github.com/ollama/ollama/issues/2693
2,149,835,982
I_kwDOJ0Z1Ps6AI-TO
2,693
Ollama windows version not working
{ "login": "sdrshn-nmbr", "id": 122426494, "node_id": "U_kgDOB0wUfg", "avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sdrshn-nmbr", "html_url": "https://github.com/sdrshn-nmbr", "followers_url": "https://api.github.com/users/sdrshn-nmbr/followers", "following_url": "https://api.github.com/users/sdrshn-nmbr/following{/other_user}", "gists_url": "https://api.github.com/users/sdrshn-nmbr/gists{/gist_id}", "starred_url": "https://api.github.com/users/sdrshn-nmbr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sdrshn-nmbr/subscriptions", "organizations_url": "https://api.github.com/users/sdrshn-nmbr/orgs", "repos_url": "https://api.github.com/users/sdrshn-nmbr/repos", "events_url": "https://api.github.com/users/sdrshn-nmbr/events{/privacy}", "received_events_url": "https://api.github.com/users/sdrshn-nmbr/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-02-22T19:55:55
2024-02-22T20:24:32
2024-02-22T20:24:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It was working fine even yesterday, but I got an update notification and it hasn't been working since. I even tried deleting and reinstalling the installer exe, but it seems the app shows up for a few seconds and then disappears again, but powershell still recognizes the command - it just says ollama not running.
{ "login": "sdrshn-nmbr", "id": 122426494, "node_id": "U_kgDOB0wUfg", "avatar_url": "https://avatars.githubusercontent.com/u/122426494?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sdrshn-nmbr", "html_url": "https://github.com/sdrshn-nmbr", "followers_url": "https://api.github.com/users/sdrshn-nmbr/followers", "following_url": "https://api.github.com/users/sdrshn-nmbr/following{/other_user}", "gists_url": "https://api.github.com/users/sdrshn-nmbr/gists{/gist_id}", "starred_url": "https://api.github.com/users/sdrshn-nmbr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sdrshn-nmbr/subscriptions", "organizations_url": "https://api.github.com/users/sdrshn-nmbr/orgs", "repos_url": "https://api.github.com/users/sdrshn-nmbr/repos", "events_url": "https://api.github.com/users/sdrshn-nmbr/events{/privacy}", "received_events_url": "https://api.github.com/users/sdrshn-nmbr/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2693/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2693/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5758
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5758/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5758/comments
https://api.github.com/repos/ollama/ollama/issues/5758/events
https://github.com/ollama/ollama/pull/5758
2,414,681,278
PR_kwDOJ0Z1Ps51sq18
5,758
marshal json automatically for some template values
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-17T22:02:46
2024-07-17T22:35:13
2024-07-17T22:35:11
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5758", "html_url": "https://github.com/ollama/ollama/pull/5758", "diff_url": "https://github.com/ollama/ollama/pull/5758.diff", "patch_url": "https://github.com/ollama/ollama/pull/5758.patch", "merged_at": "2024-07-17T22:35:11" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5758/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5758/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3678
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3678/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3678/comments
https://api.github.com/repos/ollama/ollama/issues/3678/events
https://github.com/ollama/ollama/pull/3678
2,246,645,169
PR_kwDOJ0Z1Ps5s1fe1
3,678
darwin: no partial offloading if required memory greater than system
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-16T18:23:02
2024-04-16T19:05:57
2024-04-16T19:05:57
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3678", "html_url": "https://github.com/ollama/ollama/pull/3678", "diff_url": "https://github.com/ollama/ollama/pull/3678.diff", "patch_url": "https://github.com/ollama/ollama/pull/3678.patch", "merged_at": "2024-04-16T19:05:56" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3678/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3678/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4381
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4381/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4381/comments
https://api.github.com/repos/ollama/ollama/issues/4381/events
https://github.com/ollama/ollama/issues/4381
2,291,487,916
I_kwDOJ0Z1Ps6IlVSs
4,381
Yi 1.5 was released
{ "login": "ddpasa", "id": 112642920, "node_id": "U_kgDOBrbLaA", "avatar_url": "https://avatars.githubusercontent.com/u/112642920?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ddpasa", "html_url": "https://github.com/ddpasa", "followers_url": "https://api.github.com/users/ddpasa/followers", "following_url": "https://api.github.com/users/ddpasa/following{/other_user}", "gists_url": "https://api.github.com/users/ddpasa/gists{/gist_id}", "starred_url": "https://api.github.com/users/ddpasa/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ddpasa/subscriptions", "organizations_url": "https://api.github.com/users/ddpasa/orgs", "repos_url": "https://api.github.com/users/ddpasa/repos", "events_url": "https://api.github.com/users/ddpasa/events{/privacy}", "received_events_url": "https://api.github.com/users/ddpasa/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-05-12T20:23:39
2024-05-14T03:35:51
2024-05-13T21:34:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://github.com/01-ai/Yi-1.5 https://huggingface.co/collections/01-ai/yi-15-2024-05-663f3ecab5f815a3eaca7ca8
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4381/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4381/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/783
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/783/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/783/comments
https://api.github.com/repos/ollama/ollama/issues/783/events
https://github.com/ollama/ollama/pull/783
1,942,570,643
PR_kwDOJ0Z1Ps5cxVPh
783
fix: offloading on low end GPUs
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-10-13T20:08:46
2023-10-13T21:36:45
2023-10-13T21:36:44
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/783", "html_url": "https://github.com/ollama/ollama/pull/783", "diff_url": "https://github.com/ollama/ollama/pull/783.diff", "patch_url": "https://github.com/ollama/ollama/pull/783.patch", "merged_at": "2023-10-13T21:36:44" }
Fixes two issues when using low end GPUs: GPUs with low VRAM are disproportionately affected by overhead when offloading so any device that has less than 2GB VRAM will be exclusively CPU unless overwritten by num_gpu. A CUDA-enabled runner will still offload to GPU even if num_gpu is 0. This is problematic when the GPU doesn't support a compatible version of CUDA. In this case, select the CPU runner instead. Caveat: for MacOS (darwin) `go generate` only builds Metal on ARM so it shouldn't be marked as `Accelerated` since there's no fallback
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/783/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/783/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/861
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/861/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/861/comments
https://api.github.com/repos/ollama/ollama/issues/861/events
https://github.com/ollama/ollama/pull/861
1,954,903,919
PR_kwDOJ0Z1Ps5dakP4
861
Added /clear command to clear the buffer
{ "login": "pepperoni21", "id": 29759371, "node_id": "MDQ6VXNlcjI5NzU5Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pepperoni21", "html_url": "https://github.com/pepperoni21", "followers_url": "https://api.github.com/users/pepperoni21/followers", "following_url": "https://api.github.com/users/pepperoni21/following{/other_user}", "gists_url": "https://api.github.com/users/pepperoni21/gists{/gist_id}", "starred_url": "https://api.github.com/users/pepperoni21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pepperoni21/subscriptions", "organizations_url": "https://api.github.com/users/pepperoni21/orgs", "repos_url": "https://api.github.com/users/pepperoni21/repos", "events_url": "https://api.github.com/users/pepperoni21/events{/privacy}", "received_events_url": "https://api.github.com/users/pepperoni21/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-10-20T19:25:06
2023-10-20T20:10:22
2023-10-20T20:05:19
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/861", "html_url": "https://github.com/ollama/ollama/pull/861", "diff_url": "https://github.com/ollama/ollama/pull/861.diff", "patch_url": "https://github.com/ollama/ollama/pull/861.patch", "merged_at": null }
Hey, I thought a /clear command could enhance user experience. It should work for both windows and linux/mac.
{ "login": "pepperoni21", "id": 29759371, "node_id": "MDQ6VXNlcjI5NzU5Mzcx", "avatar_url": "https://avatars.githubusercontent.com/u/29759371?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pepperoni21", "html_url": "https://github.com/pepperoni21", "followers_url": "https://api.github.com/users/pepperoni21/followers", "following_url": "https://api.github.com/users/pepperoni21/following{/other_user}", "gists_url": "https://api.github.com/users/pepperoni21/gists{/gist_id}", "starred_url": "https://api.github.com/users/pepperoni21/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pepperoni21/subscriptions", "organizations_url": "https://api.github.com/users/pepperoni21/orgs", "repos_url": "https://api.github.com/users/pepperoni21/repos", "events_url": "https://api.github.com/users/pepperoni21/events{/privacy}", "received_events_url": "https://api.github.com/users/pepperoni21/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/861/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/861/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4830
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4830/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4830/comments
https://api.github.com/repos/ollama/ollama/issues/4830/events
https://github.com/ollama/ollama/issues/4830
2,335,295,422
I_kwDOJ0Z1Ps6LMce-
4,830
model import to Ollama
{ "login": "fingerfree77", "id": 124861936, "node_id": "U_kgDOB3E98A", "avatar_url": "https://avatars.githubusercontent.com/u/124861936?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fingerfree77", "html_url": "https://github.com/fingerfree77", "followers_url": "https://api.github.com/users/fingerfree77/followers", "following_url": "https://api.github.com/users/fingerfree77/following{/other_user}", "gists_url": "https://api.github.com/users/fingerfree77/gists{/gist_id}", "starred_url": "https://api.github.com/users/fingerfree77/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fingerfree77/subscriptions", "organizations_url": "https://api.github.com/users/fingerfree77/orgs", "repos_url": "https://api.github.com/users/fingerfree77/repos", "events_url": "https://api.github.com/users/fingerfree77/events{/privacy}", "received_events_url": "https://api.github.com/users/fingerfree77/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
3
2024-06-05T08:54:55
2024-09-04T13:31:21
2024-09-04T13:31:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Is it possible to import model to Ollama from HuggingFace which is not listed in the script convert-hf-to-gguf.py? In my case model: https://huggingface.co/ai-forever/ruGPT-3.5-13B When I try import got error: ``` Traceback (most recent call last): File "llm/llama.cpp/convert-hf-to-gguf.py", line 2865, in <module> main() File "llm/llama.cpp/convert-hf-to-gguf.py", line 2850, in main model_instance.set_vocab() File "llm/llama.cpp/convert-hf-to-gguf.py", line 114, in set_vocab self._set_vocab_gpt2() File "llm/llama.cpp/convert-hf-to-gguf.py", line 500, in _set_vocab_gpt2 tokens, toktypes, tokpre = self.get_vocab_base() File "llm/llama.cpp/convert-hf-to-gguf.py", line 379, in get_vocab_base tokpre = self.get_vocab_base_pre(tokenizer) File "llm/llama.cpp/convert-hf-to-gguf.py", line 491, in get_vocab_base_pre raise NotImplementedError("BPE pre-tokenizer was not recognized - update get_vocab_base_pre()") NotImplementedError: BPE pre-tokenizer was not recognized - update get_vocab_base_pre() ```
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4830/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4830/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3569
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3569/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3569/comments
https://api.github.com/repos/ollama/ollama/issues/3569/events
https://github.com/ollama/ollama/issues/3569
2,234,581,592
I_kwDOJ0Z1Ps6FMQJY
3,569
Could anyone tell me the purpose of "ollama-linux-amd64-rocm.tgz" in the released files?
{ "login": "airforce011", "id": 36444699, "node_id": "MDQ6VXNlcjM2NDQ0Njk5", "avatar_url": "https://avatars.githubusercontent.com/u/36444699?v=4", "gravatar_id": "", "url": "https://api.github.com/users/airforce011", "html_url": "https://github.com/airforce011", "followers_url": "https://api.github.com/users/airforce011/followers", "following_url": "https://api.github.com/users/airforce011/following{/other_user}", "gists_url": "https://api.github.com/users/airforce011/gists{/gist_id}", "starred_url": "https://api.github.com/users/airforce011/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/airforce011/subscriptions", "organizations_url": "https://api.github.com/users/airforce011/orgs", "repos_url": "https://api.github.com/users/airforce011/repos", "events_url": "https://api.github.com/users/airforce011/events{/privacy}", "received_events_url": "https://api.github.com/users/airforce011/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-04-10T01:29:39
2024-04-23T00:02:08
2024-04-23T00:02:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? I'm trying to run Ollama on Linux (Ubuntu 22.04) with ROCm, but I'm unable to use it even after following the installation instructions in the documentation. Do I need to use the "ollama-linux-amd64-rocm.tgz" file? However, I'm not sure how to use this file, and there are no instructions provided. Any help would be greatly appreciated! ### How should we solve this? _No response_ ### What is the impact of not solving this? _No response_ ### Anything else? _No response_
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3569/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3569/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4963
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4963/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4963/comments
https://api.github.com/repos/ollama/ollama/issues/4963/events
https://github.com/ollama/ollama/issues/4963
2,344,382,940
I_kwDOJ0Z1Ps6LvHHc
4,963
Ollama in private EC2
{ "login": "arpitdish", "id": 123781157, "node_id": "U_kgDOB2DAJQ", "avatar_url": "https://avatars.githubusercontent.com/u/123781157?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arpitdish", "html_url": "https://github.com/arpitdish", "followers_url": "https://api.github.com/users/arpitdish/followers", "following_url": "https://api.github.com/users/arpitdish/following{/other_user}", "gists_url": "https://api.github.com/users/arpitdish/gists{/gist_id}", "starred_url": "https://api.github.com/users/arpitdish/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arpitdish/subscriptions", "organizations_url": "https://api.github.com/users/arpitdish/orgs", "repos_url": "https://api.github.com/users/arpitdish/repos", "events_url": "https://api.github.com/users/arpitdish/events{/privacy}", "received_events_url": "https://api.github.com/users/arpitdish/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-06-10T16:40:51
2024-07-10T23:01:21
2024-07-10T23:01:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? How to run ollama on private EC2 ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version _No response_
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4963/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4963/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4290
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4290/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4290/comments
https://api.github.com/repos/ollama/ollama/issues/4290/events
https://github.com/ollama/ollama/pull/4290
2,288,039,651
PR_kwDOJ0Z1Ps5vAm1e
4,290
fix(server): prune files
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-05-09T16:25:11
2024-11-23T21:28:14
2024-11-23T21:28:14
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4290", "html_url": "https://github.com/ollama/ollama/pull/4290", "diff_url": "https://github.com/ollama/ollama/pull/4290.diff", "patch_url": "https://github.com/ollama/ollama/pull/4290.patch", "merged_at": null }
follow up to #4269
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4290/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4290/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5627
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5627/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5627/comments
https://api.github.com/repos/ollama/ollama/issues/5627/events
https://github.com/ollama/ollama/issues/5627
2,402,504,053
I_kwDOJ0Z1Ps6PM011
5,627
unknown model architecture: 'gemma2'
{ "login": "mouse33333", "id": 4304174, "node_id": "MDQ6VXNlcjQzMDQxNzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/4304174?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mouse33333", "html_url": "https://github.com/mouse33333", "followers_url": "https://api.github.com/users/mouse33333/followers", "following_url": "https://api.github.com/users/mouse33333/following{/other_user}", "gists_url": "https://api.github.com/users/mouse33333/gists{/gist_id}", "starred_url": "https://api.github.com/users/mouse33333/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mouse33333/subscriptions", "organizations_url": "https://api.github.com/users/mouse33333/orgs", "repos_url": "https://api.github.com/users/mouse33333/repos", "events_url": "https://api.github.com/users/mouse33333/events{/privacy}", "received_events_url": "https://api.github.com/users/mouse33333/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-07-11T07:47:19
2024-07-11T08:02:42
2024-07-11T07:55:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm trying to run models on my internal Intel GPU. Here's the software env: 1. activate a virtual env in Anaconda 2. already installed ipex-llm[cpp], ipex-llm[xpu] 3. start the Ollama serve 4. start gemma2 model Then the log said: unknown model architecture: 'gemma2' Here's the sreenshot of the log: ![screenshot-20240711-154648](https://github.com/ollama/ollama/assets/4304174/b7e59f45-c7bc-4e62-86dc-93f1083b8b55) ### OS Windows, Docker ### GPU Intel ### CPU Intel ### Ollama version 0.2.1
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5627/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5627/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4011
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4011/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4011/comments
https://api.github.com/repos/ollama/ollama/issues/4011/events
https://github.com/ollama/ollama/issues/4011
2,267,922,429
I_kwDOJ0Z1Ps6HLb_9
4,011
When my ollama has both the gemma and llama large models, how can I enable them at the same time?Thank you.
{ "login": "joylijoy", "id": 167687386, "node_id": "U_kgDOCf602g", "avatar_url": "https://avatars.githubusercontent.com/u/167687386?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joylijoy", "html_url": "https://github.com/joylijoy", "followers_url": "https://api.github.com/users/joylijoy/followers", "following_url": "https://api.github.com/users/joylijoy/following{/other_user}", "gists_url": "https://api.github.com/users/joylijoy/gists{/gist_id}", "starred_url": "https://api.github.com/users/joylijoy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joylijoy/subscriptions", "organizations_url": "https://api.github.com/users/joylijoy/orgs", "repos_url": "https://api.github.com/users/joylijoy/repos", "events_url": "https://api.github.com/users/joylijoy/events{/privacy}", "received_events_url": "https://api.github.com/users/joylijoy/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-04-29T00:54:44
2024-05-01T04:56:32
2024-05-01T04:56:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When my ollama has both the gemma and llama large models, how can I enable them at the same time? Does pressing CTRL+D mean to exit the large model? Thank you.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4011/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4011/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8457
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8457/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8457/comments
https://api.github.com/repos/ollama/ollama/issues/8457/events
https://github.com/ollama/ollama/issues/8457
2,792,336,253
I_kwDOJ0Z1Ps6mb6t9
8,457
model wanted in ollama please:Qwen2.5-Math-PRM-7B
{ "login": "twythebest", "id": 89891289, "node_id": "MDQ6VXNlcjg5ODkxMjg5", "avatar_url": "https://avatars.githubusercontent.com/u/89891289?v=4", "gravatar_id": "", "url": "https://api.github.com/users/twythebest", "html_url": "https://github.com/twythebest", "followers_url": "https://api.github.com/users/twythebest/followers", "following_url": "https://api.github.com/users/twythebest/following{/other_user}", "gists_url": "https://api.github.com/users/twythebest/gists{/gist_id}", "starred_url": "https://api.github.com/users/twythebest/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/twythebest/subscriptions", "organizations_url": "https://api.github.com/users/twythebest/orgs", "repos_url": "https://api.github.com/users/twythebest/repos", "events_url": "https://api.github.com/users/twythebest/events{/privacy}", "received_events_url": "https://api.github.com/users/twythebest/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2025-01-16T10:36:00
2025-01-16T10:36:00
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
model wanted in ollama please:Qwen2.5-Math-PRM-7B
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8457/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8457/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6544
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6544/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6544/comments
https://api.github.com/repos/ollama/ollama/issues/6544/events
https://github.com/ollama/ollama/issues/6544
2,492,824,185
I_kwDOJ0Z1Ps6UlXp5
6,544
Specifying options via openai client extra_body are not handled by ollama
{ "login": "gaardhus", "id": 46934916, "node_id": "MDQ6VXNlcjQ2OTM0OTE2", "avatar_url": "https://avatars.githubusercontent.com/u/46934916?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gaardhus", "html_url": "https://github.com/gaardhus", "followers_url": "https://api.github.com/users/gaardhus/followers", "following_url": "https://api.github.com/users/gaardhus/following{/other_user}", "gists_url": "https://api.github.com/users/gaardhus/gists{/gist_id}", "starred_url": "https://api.github.com/users/gaardhus/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gaardhus/subscriptions", "organizations_url": "https://api.github.com/users/gaardhus/orgs", "repos_url": "https://api.github.com/users/gaardhus/repos", "events_url": "https://api.github.com/users/gaardhus/events{/privacy}", "received_events_url": "https://api.github.com/users/gaardhus/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
5
2024-08-28T19:19:11
2024-09-05T21:47:15
2024-09-04T13:25:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? So I've been trying to set the num_ctx for the mistral-nemo through the openai api client, however it does not seem to have an effect. ```python # Ollama client client = OllamaAsyncClient(host=base_url, **kwargs) chat_completion = await client.chat( messages=messages, model=model, stream=stream, options={ "num_ctx": 64_000, "temperature": temperature, "stop": stop_tokens, }, **kwargs, ) message = chat_completion["message"]["content"].strip() # OpenAI client client = AsyncOpenAI(api_key=api_key, base_url=base_url, **kwargs) chat_completion = await client.chat.completions.create( messages=messages, model=model, temperature=temperature, stream=stream, stop=stop_tokens, extra_body={"options": {"num_ctx": 64_000}}, **kwargs, ) message = chat_completion.choices[0].message.content.strip() ``` Doing it with ollama works, but the extra_body argument from openai seems to be handled as an extra field rather than merged with the rest of the request. ``` 'model': 'mistral-nemo:12b-instruct-2407-q8_0', 'stop': None, 'stream': False, 'temperature': 0.7}, 'extra_json': {'options': {'num_ctx': 64000}}} ``` I guess the solution would be to unpack the extra_json field on the server end? ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.8
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6544/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6544/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7106
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7106/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7106/comments
https://api.github.com/repos/ollama/ollama/issues/7106/events
https://github.com/ollama/ollama/issues/7106
2,568,514,399
I_kwDOJ0Z1Ps6ZGGtf
7,106
Update /api/show endpoint to return a boolean describing if the model supports FIM
{ "login": "tomasz-stefaniak", "id": 7911227, "node_id": "MDQ6VXNlcjc5MTEyMjc=", "avatar_url": "https://avatars.githubusercontent.com/u/7911227?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tomasz-stefaniak", "html_url": "https://github.com/tomasz-stefaniak", "followers_url": "https://api.github.com/users/tomasz-stefaniak/followers", "following_url": "https://api.github.com/users/tomasz-stefaniak/following{/other_user}", "gists_url": "https://api.github.com/users/tomasz-stefaniak/gists{/gist_id}", "starred_url": "https://api.github.com/users/tomasz-stefaniak/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tomasz-stefaniak/subscriptions", "organizations_url": "https://api.github.com/users/tomasz-stefaniak/orgs", "repos_url": "https://api.github.com/users/tomasz-stefaniak/repos", "events_url": "https://api.github.com/users/tomasz-stefaniak/events{/privacy}", "received_events_url": "https://api.github.com/users/tomasz-stefaniak/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-10-06T07:32:55
2024-11-17T14:49:48
2024-11-17T14:49:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, We support Ollama as a [model provider in Continue.dev](https://docs.continue.dev/customize/model-providers/ollama) and have recently added FIM (Fill-in-the-Middle) support. Currently, there is no API we are aware of that can be used to determine if a model supports FIM, so we guess its capabilities by checking for a ".Suffix" variable in the template returned from Ollama's `/api/show` endpoint. To improve accuracy and stability, we would appreciate receiving this information directly as a boolean value from the `/api/show` endpoint. This would allow us to integrate FIM support more reliably and provide better service to Continue users. Context: [GitHub Link](https://github.com/continuedev/continue/pull/2452/files#diff-3f4bdaca0060eb741057719bcc865486ab770e4ea1659b317f77235adc887c37R81)
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7106/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7106/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/6749
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6749/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6749/comments
https://api.github.com/repos/ollama/ollama/issues/6749/events
https://github.com/ollama/ollama/issues/6749
2,519,216,246
I_kwDOJ0Z1Ps6WKDB2
6,749
Add version when the docker container is starting
{ "login": "svaningelgem", "id": 199434, "node_id": "MDQ6VXNlcjE5OTQzNA==", "avatar_url": "https://avatars.githubusercontent.com/u/199434?v=4", "gravatar_id": "", "url": "https://api.github.com/users/svaningelgem", "html_url": "https://github.com/svaningelgem", "followers_url": "https://api.github.com/users/svaningelgem/followers", "following_url": "https://api.github.com/users/svaningelgem/following{/other_user}", "gists_url": "https://api.github.com/users/svaningelgem/gists{/gist_id}", "starred_url": "https://api.github.com/users/svaningelgem/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/svaningelgem/subscriptions", "organizations_url": "https://api.github.com/users/svaningelgem/orgs", "repos_url": "https://api.github.com/users/svaningelgem/repos", "events_url": "https://api.github.com/users/svaningelgem/events{/privacy}", "received_events_url": "https://api.github.com/users/svaningelgem/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-09-11T09:46:48
2024-09-12T03:50:22
2024-09-11T23:56:59
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It'd be nice to know what version is running when sharing logs. So my feature request is to add the current version when you start the docker container.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6749/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6749/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4271
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4271/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4271/comments
https://api.github.com/repos/ollama/ollama/issues/4271/events
https://github.com/ollama/ollama/issues/4271
2,286,720,880
I_kwDOJ0Z1Ps6ITJdw
4,271
Partial pruning does not wrok
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
0
2024-05-09T01:17:56
2024-05-09T01:17:56
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ``` time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:cfcf93119280c4a10c1df57335bad341e000cabbc4faff125531d941a5b0befa:partial:13': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:cfcf93119280c4a10c1df57335bad341e000cabbc4faff125531d941a5b0befa:partial:67': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:4fe022a8902336d3c452c88f7aca5590f5b5b02ccfd06320fdefab02412e1f0b:partial:18': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:431800fe7a30f378018e76cdec8ed5f99e6f81b22fc5f63b0486e326f0b25545:partial:37': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:4fe022a8902336d3c452c88f7aca5590f5b5b02ccfd06320fdefab02412e1f0b:partial:23': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:660 msg="couldn't get file path for 'sha256:4fe022a8902336d3c452c88f7aca5590f5b5b02ccfd06320fdefab02412e1f0b:partial:8': invalid digest format" time=2024-05-08T21:17:41.854-04:00 level=INFO source=images.go:704 msg="total unused blobs removed: 190" ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4271/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4271/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8649
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8649/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8649/comments
https://api.github.com/repos/ollama/ollama/issues/8649/events
https://github.com/ollama/ollama/issues/8649
2,817,207,537
I_kwDOJ0Z1Ps6n6yzx
8,649
Short run response duration calculations are off
{ "login": "NerdyShawn", "id": 16088118, "node_id": "MDQ6VXNlcjE2MDg4MTE4", "avatar_url": "https://avatars.githubusercontent.com/u/16088118?v=4", "gravatar_id": "", "url": "https://api.github.com/users/NerdyShawn", "html_url": "https://github.com/NerdyShawn", "followers_url": "https://api.github.com/users/NerdyShawn/followers", "following_url": "https://api.github.com/users/NerdyShawn/following{/other_user}", "gists_url": "https://api.github.com/users/NerdyShawn/gists{/gist_id}", "starred_url": "https://api.github.com/users/NerdyShawn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NerdyShawn/subscriptions", "organizations_url": "https://api.github.com/users/NerdyShawn/orgs", "repos_url": "https://api.github.com/users/NerdyShawn/repos", "events_url": "https://api.github.com/users/NerdyShawn/events{/privacy}", "received_events_url": "https://api.github.com/users/NerdyShawn/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2025-01-29T04:22:55
2025-01-29T14:03:03
2025-01-29T14:03:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Running the smaller `deepseek-r1:1.5b` model it seems like the very short time and duration how it is being calculated is off given the duration in the response. It seems since its close to zero it throws off how the time is measured. --- ![Image](https://github.com/user-attachments/assets/aba179e4-7c51-4c44-8503-2f91e60809a2) --- ```json date && time curl -s https://ollama.somecooldomain.lan/api/generate -d '{ "model": "deepseek-r1:1.5b", "prompt": "What is the meaning of life?", "stream": false }' | jq Tue Jan 28 11:10:42 PM EST 2025 { "model": "deepseek-r1:1.5b", "created_at": "2025-01-29T04:10:42.569719236Z", "response": "<think>\n\n</think>\n\nI am sorry, I cannot answer that question.", "done": true, "done_reason": "stop", "context": [ 151644, 3838, 374, 279, 7290, 315, 2272, 30, 151645, 151648, 271, 151649, 271, 40, 1079, 14589, 11, 358, 4157, 4226, 429, 3405, 13 ], "total_duration": 256103089, "load_duration": 144586102, "prompt_eval_count": 10, "prompt_eval_duration": 24000000, "eval_count": 15, "eval_duration": 86000000 } real 0m0.345s user 0m0.050s sys 0m0.016s ``` The real time it took to execute was under a half second but the durations are all wrong. ### OS Docker ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.6-0-g2539f2d-dirty
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8649/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8649/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2864
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2864/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2864/comments
https://api.github.com/repos/ollama/ollama/issues/2864/events
https://github.com/ollama/ollama/issues/2864
2,163,502,283
I_kwDOJ0Z1Ps6A9GzL
2,864
Read and write files (To allow for more types of AI models in Ollama)
{ "login": "trymeouteh", "id": 31172274, "node_id": "MDQ6VXNlcjMxMTcyMjc0", "avatar_url": "https://avatars.githubusercontent.com/u/31172274?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trymeouteh", "html_url": "https://github.com/trymeouteh", "followers_url": "https://api.github.com/users/trymeouteh/followers", "following_url": "https://api.github.com/users/trymeouteh/following{/other_user}", "gists_url": "https://api.github.com/users/trymeouteh/gists{/gist_id}", "starred_url": "https://api.github.com/users/trymeouteh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/trymeouteh/subscriptions", "organizations_url": "https://api.github.com/users/trymeouteh/orgs", "repos_url": "https://api.github.com/users/trymeouteh/repos", "events_url": "https://api.github.com/users/trymeouteh/events{/privacy}", "received_events_url": "https://api.github.com/users/trymeouteh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-03-01T14:12:41
2024-03-08T14:09:08
2024-03-07T18:54:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
To my understanding it will seem Ollama can only support text chat based models and to my understand it could support more kinds of models by adding a reading files and writing files feature... - A way to have a model read one or more file on your system and be able to send it to the AI model to process. - A way for the AI model to return a file, by generating a file (Writing a file) onto the system and them being able to link it into the chat response. To my understanding, this can allow Ollama to work with more model types such as text to image models, image to image models, text to audio models, audio to text models, audio to audio models, etc.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2864/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2864/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/502
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/502/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/502/comments
https://api.github.com/repos/ollama/ollama/issues/502/events
https://github.com/ollama/ollama/issues/502
1,889,051,436
I_kwDOJ0Z1Ps5wmKMs
502
Better Document /api/embeddings
{ "login": "FairyTail2000", "id": 22645621, "node_id": "MDQ6VXNlcjIyNjQ1NjIx", "avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FairyTail2000", "html_url": "https://github.com/FairyTail2000", "followers_url": "https://api.github.com/users/FairyTail2000/followers", "following_url": "https://api.github.com/users/FairyTail2000/following{/other_user}", "gists_url": "https://api.github.com/users/FairyTail2000/gists{/gist_id}", "starred_url": "https://api.github.com/users/FairyTail2000/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/FairyTail2000/subscriptions", "organizations_url": "https://api.github.com/users/FairyTail2000/orgs", "repos_url": "https://api.github.com/users/FairyTail2000/repos", "events_url": "https://api.github.com/users/FairyTail2000/events{/privacy}", "received_events_url": "https://api.github.com/users/FairyTail2000/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2023-09-10T08:37:43
2023-10-26T00:12:03
2023-10-26T00:12:03
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It's currently not very clear on how to use is and how it's relation is to the EMBED Modelfile instruction. Can you enlighten me and/or update the wiki and examples?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/502/reactions", "total_count": 4, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 2 }
https://api.github.com/repos/ollama/ollama/issues/502/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7634
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7634/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7634/comments
https://api.github.com/repos/ollama/ollama/issues/7634/events
https://github.com/ollama/ollama/issues/7634
2,653,045,312
I_kwDOJ0Z1Ps6eIkJA
7,634
Custom tool are not getting called in tool call
{ "login": "Vikneshkumarmohan", "id": 71374469, "node_id": "MDQ6VXNlcjcxMzc0NDY5", "avatar_url": "https://avatars.githubusercontent.com/u/71374469?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Vikneshkumarmohan", "html_url": "https://github.com/Vikneshkumarmohan", "followers_url": "https://api.github.com/users/Vikneshkumarmohan/followers", "following_url": "https://api.github.com/users/Vikneshkumarmohan/following{/other_user}", "gists_url": "https://api.github.com/users/Vikneshkumarmohan/gists{/gist_id}", "starred_url": "https://api.github.com/users/Vikneshkumarmohan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Vikneshkumarmohan/subscriptions", "organizations_url": "https://api.github.com/users/Vikneshkumarmohan/orgs", "repos_url": "https://api.github.com/users/Vikneshkumarmohan/repos", "events_url": "https://api.github.com/users/Vikneshkumarmohan/events{/privacy}", "received_events_url": "https://api.github.com/users/Vikneshkumarmohan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-11-12T18:51:44
2024-12-03T03:55:02
2024-12-02T08:01:16
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? when using ChatOllama in Langgraph tool call is not happening for the custom tools , but it is working as expected with tools like tavily search , https://langchain-ai.github.io/langgraph/tutorials/multi_agent/multi-agent-collaboration/#invoke above is the example for reference code , when working with ollama LLM the custom tool is not calling (custom tool : python_repl), ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version Llama3.2
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7634/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/7634/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6186
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6186/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6186/comments
https://api.github.com/repos/ollama/ollama/issues/6186/events
https://github.com/ollama/ollama/pull/6186
2,449,377,760
PR_kwDOJ0Z1Ps53fc2u
6,186
Implement linux NUMA detection
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-08-05T20:01:08
2024-08-05T22:20:11
2024-08-05T22:20:07
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6186", "html_url": "https://github.com/ollama/ollama/pull/6186", "diff_url": "https://github.com/ollama/ollama/pull/6186.diff", "patch_url": "https://github.com/ollama/ollama/pull/6186.patch", "merged_at": "2024-08-05T22:20:07" }
If the system has multiple numa nodes, enable numa support in llama.cpp If we detect `numactl` in the path, use that, else use the basic "distribute" mode. This also removes the `use_numa` setting as the bool type is no longer useful given this parameter now requires a mode. Fixes #6093 Fixes #2496
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6186/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6186/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5715
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5715/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5715/comments
https://api.github.com/repos/ollama/ollama/issues/5715/events
https://github.com/ollama/ollama/issues/5715
2,410,125,004
I_kwDOJ0Z1Ps6Pp5bM
5,715
Allow using `"""` in TEMPLATE Modelfile command
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-07-16T03:33:31
2024-07-16T03:33:31
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5715/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5715/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7121
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7121/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7121/comments
https://api.github.com/repos/ollama/ollama/issues/7121/events
https://github.com/ollama/ollama/pull/7121
2,571,261,705
PR_kwDOJ0Z1Ps5925z4
7,121
llama: refine developer docs
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-07T19:18:10
2024-10-07T19:43:49
2024-10-07T19:43:46
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7121", "html_url": "https://github.com/ollama/ollama/pull/7121", "diff_url": "https://github.com/ollama/ollama/pull/7121.diff", "patch_url": "https://github.com/ollama/ollama/pull/7121.patch", "merged_at": "2024-10-07T19:43:46" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7121/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7121/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8183
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8183/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8183/comments
https://api.github.com/repos/ollama/ollama/issues/8183/events
https://github.com/ollama/ollama/issues/8183
2,752,749,059
I_kwDOJ0Z1Ps6kE54D
8,183
How do I specify specific GPUs when running a model?
{ "login": "any35", "id": 2082210, "node_id": "MDQ6VXNlcjIwODIyMTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2082210?v=4", "gravatar_id": "", "url": "https://api.github.com/users/any35", "html_url": "https://github.com/any35", "followers_url": "https://api.github.com/users/any35/followers", "following_url": "https://api.github.com/users/any35/following{/other_user}", "gists_url": "https://api.github.com/users/any35/gists{/gist_id}", "starred_url": "https://api.github.com/users/any35/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/any35/subscriptions", "organizations_url": "https://api.github.com/users/any35/orgs", "repos_url": "https://api.github.com/users/any35/repos", "events_url": "https://api.github.com/users/any35/events{/privacy}", "received_events_url": "https://api.github.com/users/any35/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-12-20T13:30:03
2025-01-13T01:43:05
2025-01-13T01:43:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have a server with 16 RTX4090 graphics cards installed. Now I need to run different models at the same time and specify the corresponding GPUs. How can I do this? For example: run llama 3.3 on GPU0,1; llama 3.2 on gpu 2;...
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8183/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8183/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1306
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1306/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1306/comments
https://api.github.com/repos/ollama/ollama/issues/1306/events
https://github.com/ollama/ollama/pull/1306
2,014,904,973
PR_kwDOJ0Z1Ps5glFGZ
1,306
Fix Docker image not using GPU
{ "login": "bwest2397", "id": 73724178, "node_id": "MDQ6VXNlcjczNzI0MTc4", "avatar_url": "https://avatars.githubusercontent.com/u/73724178?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bwest2397", "html_url": "https://github.com/bwest2397", "followers_url": "https://api.github.com/users/bwest2397/followers", "following_url": "https://api.github.com/users/bwest2397/following{/other_user}", "gists_url": "https://api.github.com/users/bwest2397/gists{/gist_id}", "starred_url": "https://api.github.com/users/bwest2397/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bwest2397/subscriptions", "organizations_url": "https://api.github.com/users/bwest2397/orgs", "repos_url": "https://api.github.com/users/bwest2397/repos", "events_url": "https://api.github.com/users/bwest2397/events{/privacy}", "received_events_url": "https://api.github.com/users/bwest2397/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
11
2023-11-28T17:06:07
2023-12-12T19:34:56
2023-11-28T21:25:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1306", "html_url": "https://github.com/ollama/ollama/pull/1306", "diff_url": "https://github.com/ollama/ollama/pull/1306.diff", "patch_url": "https://github.com/ollama/ollama/pull/1306.patch", "merged_at": null }
As it currently stands, the Docker image that gets built is seemingly unable to use the GPU despite the initial "build" stage of the Dockerfile ostensibly being built with CUDA support (i.e. built using the `nvidia/cuda` base image). As reported in https://github.com/jmorganca/ollama/issues/797, it seems that simply setting the second stage of the Dockerfile to use this same `nvidia/cuda` base image resolves the problem.
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1306/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1306/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5667
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5667/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5667/comments
https://api.github.com/repos/ollama/ollama/issues/5667/events
https://github.com/ollama/ollama/pull/5667
2,406,750,057
PR_kwDOJ0Z1Ps51SJUt
5,667
Enhance list command
{ "login": "kaichen", "id": 4048, "node_id": "MDQ6VXNlcjQwNDg=", "avatar_url": "https://avatars.githubusercontent.com/u/4048?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kaichen", "html_url": "https://github.com/kaichen", "followers_url": "https://api.github.com/users/kaichen/followers", "following_url": "https://api.github.com/users/kaichen/following{/other_user}", "gists_url": "https://api.github.com/users/kaichen/gists{/gist_id}", "starred_url": "https://api.github.com/users/kaichen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kaichen/subscriptions", "organizations_url": "https://api.github.com/users/kaichen/orgs", "repos_url": "https://api.github.com/users/kaichen/repos", "events_url": "https://api.github.com/users/kaichen/events{/privacy}", "received_events_url": "https://api.github.com/users/kaichen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-07-13T07:21:13
2024-11-22T00:49:46
2024-11-22T00:49:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5667", "html_url": "https://github.com/ollama/ollama/pull/5667", "diff_url": "https://github.com/ollama/ollama/pull/5667.diff", "patch_url": "https://github.com/ollama/ollama/pull/5667.patch", "merged_at": null }
changes - add missing filter description for list command - add sort option to list command
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5667/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5667/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/24
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/24/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/24/comments
https://api.github.com/repos/ollama/ollama/issues/24/events
https://github.com/ollama/ollama/pull/24
1,781,626,501
PR_kwDOJ0Z1Ps5USMRS
24
remove add cmd
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-06-29T23:20:02
2023-06-29T23:29:30
2023-06-29T23:29:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/24", "html_url": "https://github.com/ollama/ollama/pull/24", "diff_url": "https://github.com/ollama/ollama/pull/24.diff", "patch_url": "https://github.com/ollama/ollama/pull/24.patch", "merged_at": "2023-06-29T23:29:23" }
🧹
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/24/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/24/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2417
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2417/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2417/comments
https://api.github.com/repos/ollama/ollama/issues/2417/events
https://github.com/ollama/ollama/issues/2417
2,126,159,957
I_kwDOJ0Z1Ps5-uqBV
2,417
Tools support in ChatCompletion endpoints
{ "login": "pamelafox", "id": 297042, "node_id": "MDQ6VXNlcjI5NzA0Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/297042?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pamelafox", "html_url": "https://github.com/pamelafox", "followers_url": "https://api.github.com/users/pamelafox/followers", "following_url": "https://api.github.com/users/pamelafox/following{/other_user}", "gists_url": "https://api.github.com/users/pamelafox/gists{/gist_id}", "starred_url": "https://api.github.com/users/pamelafox/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pamelafox/subscriptions", "organizations_url": "https://api.github.com/users/pamelafox/orgs", "repos_url": "https://api.github.com/users/pamelafox/repos", "events_url": "https://api.github.com/users/pamelafox/events{/privacy}", "received_events_url": "https://api.github.com/users/pamelafox/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6657611864, "node_id": "LA_kwDOJ0Z1Ps8AAAABjNMYWA", "url": "https://api.github.com/repos/ollama/ollama/labels/compatibility", "name": "compatibility", "color": "bfdadc", "default": false, "description": "" } ]
closed
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
10
2024-02-08T22:17:27
2024-06-25T04:30:24
2024-06-25T04:30:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
We'd love tools support so we can use ollama with our existing OpenAI-using apps. Not sure if that's possible across the board with all models.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2417/reactions", "total_count": 40, "+1": 40, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2417/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3265
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3265/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3265/comments
https://api.github.com/repos/ollama/ollama/issues/3265/events
https://github.com/ollama/ollama/issues/3265
2,197,028,474
I_kwDOJ0Z1Ps6C8_56
3,265
Does ollama also plan to support the sound models?
{ "login": "insooneelife", "id": 8437769, "node_id": "MDQ6VXNlcjg0Mzc3Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/8437769?v=4", "gravatar_id": "", "url": "https://api.github.com/users/insooneelife", "html_url": "https://github.com/insooneelife", "followers_url": "https://api.github.com/users/insooneelife/followers", "following_url": "https://api.github.com/users/insooneelife/following{/other_user}", "gists_url": "https://api.github.com/users/insooneelife/gists{/gist_id}", "starred_url": "https://api.github.com/users/insooneelife/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/insooneelife/subscriptions", "organizations_url": "https://api.github.com/users/insooneelife/orgs", "repos_url": "https://api.github.com/users/insooneelife/repos", "events_url": "https://api.github.com/users/insooneelife/events{/privacy}", "received_events_url": "https://api.github.com/users/insooneelife/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
2
2024-03-20T09:18:59
2024-12-25T15:57:35
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? Currently, ollama supports most llms and I know that it also supports vision model. I thought it would be nice if the sound model could also be used through ollama. I wonder if there are any plans like this. ### How should we solve this? I know that there are currently sound models released on huggingface. There are several TTS and STT models released as open source. I think these models could also support model serving and API like llm. ### What is the impact of not solving this? _No response_ ### Anything else? _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3265/reactions", "total_count": 7, "+1": 7, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3265/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6098
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6098/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6098/comments
https://api.github.com/repos/ollama/ollama/issues/6098/events
https://github.com/ollama/ollama/issues/6098
2,440,088,309
I_kwDOJ0Z1Ps6RcMr1
6,098
Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest.
{ "login": "fanjikang", "id": 146172937, "node_id": "U_kgDOCLZsCQ", "avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fanjikang", "html_url": "https://github.com/fanjikang", "followers_url": "https://api.github.com/users/fanjikang/followers", "following_url": "https://api.github.com/users/fanjikang/following{/other_user}", "gists_url": "https://api.github.com/users/fanjikang/gists{/gist_id}", "starred_url": "https://api.github.com/users/fanjikang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fanjikang/subscriptions", "organizations_url": "https://api.github.com/users/fanjikang/orgs", "repos_url": "https://api.github.com/users/fanjikang/repos", "events_url": "https://api.github.com/users/fanjikang/events{/privacy}", "received_events_url": "https://api.github.com/users/fanjikang/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-07-31T13:38:24
2024-08-27T10:25:25
2024-08-27T10:25:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Why is the llama3 model missing after I restart Ollama? When I run “ollama run llama3”, it re-pulls the manifest. ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version v0.2.8
{ "login": "fanjikang", "id": 146172937, "node_id": "U_kgDOCLZsCQ", "avatar_url": "https://avatars.githubusercontent.com/u/146172937?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fanjikang", "html_url": "https://github.com/fanjikang", "followers_url": "https://api.github.com/users/fanjikang/followers", "following_url": "https://api.github.com/users/fanjikang/following{/other_user}", "gists_url": "https://api.github.com/users/fanjikang/gists{/gist_id}", "starred_url": "https://api.github.com/users/fanjikang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fanjikang/subscriptions", "organizations_url": "https://api.github.com/users/fanjikang/orgs", "repos_url": "https://api.github.com/users/fanjikang/repos", "events_url": "https://api.github.com/users/fanjikang/events{/privacy}", "received_events_url": "https://api.github.com/users/fanjikang/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6098/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6098/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7150
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7150/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7150/comments
https://api.github.com/repos/ollama/ollama/issues/7150/events
https://github.com/ollama/ollama/issues/7150
2,576,056,905
I_kwDOJ0Z1Ps6Zi4JJ
7,150
CORS (Cross-Origin Resource Sharing)
{ "login": "Apex-Predator-786", "id": 94287474, "node_id": "U_kgDOBZ62cg", "avatar_url": "https://avatars.githubusercontent.com/u/94287474?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Apex-Predator-786", "html_url": "https://github.com/Apex-Predator-786", "followers_url": "https://api.github.com/users/Apex-Predator-786/followers", "following_url": "https://api.github.com/users/Apex-Predator-786/following{/other_user}", "gists_url": "https://api.github.com/users/Apex-Predator-786/gists{/gist_id}", "starred_url": "https://api.github.com/users/Apex-Predator-786/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Apex-Predator-786/subscriptions", "organizations_url": "https://api.github.com/users/Apex-Predator-786/orgs", "repos_url": "https://api.github.com/users/Apex-Predator-786/repos", "events_url": "https://api.github.com/users/Apex-Predator-786/events{/privacy}", "received_events_url": "https://api.github.com/users/Apex-Predator-786/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q", "url": "https://api.github.com/repos/ollama/ollama/labels/api", "name": "api", "color": "bfdadc", "default": false, "description": "" } ]
open
false
null
[]
null
0
2024-10-09T14:08:10
2024-11-06T00:01:14
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? please enable CORS (Cross-Origin Resource Sharing) in rest api
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7150/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7150/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2771
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2771/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2771/comments
https://api.github.com/repos/ollama/ollama/issues/2771/events
https://github.com/ollama/ollama/pull/2771
2,155,434,878
PR_kwDOJ0Z1Ps5n-7UH
2,771
Bump llama.cpp to b2276
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-27T00:51:50
2024-02-27T19:29:56
2024-02-27T19:29:53
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2771", "html_url": "https://github.com/ollama/ollama/pull/2771", "diff_url": "https://github.com/ollama/ollama/pull/2771.diff", "patch_url": "https://github.com/ollama/ollama/pull/2771.patch", "merged_at": "2024-02-27T19:29:53" }
Fixes #2758
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2771/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2771/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3271
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3271/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3271/comments
https://api.github.com/repos/ollama/ollama/issues/3271/events
https://github.com/ollama/ollama/issues/3271
2,197,650,730
I_kwDOJ0Z1Ps6C_X0q
3,271
Cloudflare Tunnels Not Work After Update 1.2.9
{ "login": "projetosfsi", "id": 117665999, "node_id": "U_kgDOBwNwzw", "avatar_url": "https://avatars.githubusercontent.com/u/117665999?v=4", "gravatar_id": "", "url": "https://api.github.com/users/projetosfsi", "html_url": "https://github.com/projetosfsi", "followers_url": "https://api.github.com/users/projetosfsi/followers", "following_url": "https://api.github.com/users/projetosfsi/following{/other_user}", "gists_url": "https://api.github.com/users/projetosfsi/gists{/gist_id}", "starred_url": "https://api.github.com/users/projetosfsi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/projetosfsi/subscriptions", "organizations_url": "https://api.github.com/users/projetosfsi/orgs", "repos_url": "https://api.github.com/users/projetosfsi/repos", "events_url": "https://api.github.com/users/projetosfsi/events{/privacy}", "received_events_url": "https://api.github.com/users/projetosfsi/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
9
2024-03-20T14:15:13
2024-11-07T18:46:03
2024-03-20T16:49:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm using Cloudfalre Tunnels to connect Ollama to a web domain. Everything was working fine until the update to version 1.2.9 I started getting the 403 Access Denied error. Regardless of how I run it, with administrative access or as a regular user, nothing changes. After downgrading to version 1.2.8 everything worked normally again. To make sure the problem wasn't with Cloudflare, I noticed that when I changed the routing to another application it opened normally. ### What did you expect to see? The Ollama Api should be working on the domain of my choice as it was before ollama.mydomain.com ### Steps to reproduce Step 01: Configure Cloudflare Tunnel https://www.youtube.com/watch?v=ey4u7OUAF3c Step 02: Open API Endpoint yoursubdomain.domain.com.br In version 1.2.8 working normally In version 1.2.9 Not Found - Erro 403 ### Are there any recent changes that introduced the issue? Update Version 1.2.8 > 1.2.9 ### OS Windows ### Architecture x86 ### Platform _No response_ ### Ollama version 1.2.9 ### GPU Nvidia ### GPU info +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 551.76 Driver Version: 551.76 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name TCC/WDDM | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce RTX 3070 Ti WDDM | 00000000:06:00.0 On | N/A | | 0% 37C P8 23W / 290W | 7284MiB / 8192MiB | 6% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ ### CPU AMD ### Other software _No response_
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3271/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3271/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1730
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1730/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1730/comments
https://api.github.com/repos/ollama/ollama/issues/1730/events
https://github.com/ollama/ollama/issues/1730
2,057,739,752
I_kwDOJ0Z1Ps56pp3o
1,730
MLX backend
{ "login": "ageorgios", "id": 3126351, "node_id": "MDQ6VXNlcjMxMjYzNTE=", "avatar_url": "https://avatars.githubusercontent.com/u/3126351?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ageorgios", "html_url": "https://github.com/ageorgios", "followers_url": "https://api.github.com/users/ageorgios/followers", "following_url": "https://api.github.com/users/ageorgios/following{/other_user}", "gists_url": "https://api.github.com/users/ageorgios/gists{/gist_id}", "starred_url": "https://api.github.com/users/ageorgios/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ageorgios/subscriptions", "organizations_url": "https://api.github.com/users/ageorgios/orgs", "repos_url": "https://api.github.com/users/ageorgios/repos", "events_url": "https://api.github.com/users/ageorgios/events{/privacy}", "received_events_url": "https://api.github.com/users/ageorgios/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
57
2023-12-27T20:10:50
2025-01-24T14:12:55
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Can ollama be converted to use MLX from Apple as backend for the models ?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1730/reactions", "total_count": 204, "+1": 164, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 40, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1730/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4407
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4407/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4407/comments
https://api.github.com/repos/ollama/ollama/issues/4407/events
https://github.com/ollama/ollama/issues/4407
2,293,538,684
I_kwDOJ0Z1Ps6ItJ98
4,407
Discrepancy between FAQ and program
{ "login": "FrederikSchack", "id": 47929714, "node_id": "MDQ6VXNlcjQ3OTI5NzE0", "avatar_url": "https://avatars.githubusercontent.com/u/47929714?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FrederikSchack", "html_url": "https://github.com/FrederikSchack", "followers_url": "https://api.github.com/users/FrederikSchack/followers", "following_url": "https://api.github.com/users/FrederikSchack/following{/other_user}", "gists_url": "https://api.github.com/users/FrederikSchack/gists{/gist_id}", "starred_url": "https://api.github.com/users/FrederikSchack/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/FrederikSchack/subscriptions", "organizations_url": "https://api.github.com/users/FrederikSchack/orgs", "repos_url": "https://api.github.com/users/FrederikSchack/repos", "events_url": "https://api.github.com/users/FrederikSchack/events{/privacy}", "received_events_url": "https://api.github.com/users/FrederikSchack/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-05-13T18:34:20
2024-05-14T01:02:27
2024-05-14T01:02:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? According to your FAQ, there should be a control panel in the Windows version for setting the environment variables: https://github.com/ollama/ollama/blob/main/docs/faq.md There isn't any control panel when you start the program, no apparent control panel or configuration file in the Ollama folder (\AppData\Local\Programs\ollama). ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.37
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4407/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4407/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/852
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/852/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/852/comments
https://api.github.com/repos/ollama/ollama/issues/852/events
https://github.com/ollama/ollama/issues/852
1,954,491,700
I_kwDOJ0Z1Ps50fy00
852
Compress with xz for macOS and Linux GitHub Releases
{ "login": "coolaj86", "id": 122831, "node_id": "MDQ6VXNlcjEyMjgzMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/122831?v=4", "gravatar_id": "", "url": "https://api.github.com/users/coolaj86", "html_url": "https://github.com/coolaj86", "followers_url": "https://api.github.com/users/coolaj86/followers", "following_url": "https://api.github.com/users/coolaj86/following{/other_user}", "gists_url": "https://api.github.com/users/coolaj86/gists{/gist_id}", "starred_url": "https://api.github.com/users/coolaj86/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/coolaj86/subscriptions", "organizations_url": "https://api.github.com/users/coolaj86/orgs", "repos_url": "https://api.github.com/users/coolaj86/repos", "events_url": "https://api.github.com/users/coolaj86/events{/privacy}", "received_events_url": "https://api.github.com/users/coolaj86/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-10-20T15:08:55
2023-10-26T05:14:35
2023-10-26T00:49:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I am willing to help with this. ## Why `xz`? `xz` is a standard compression tool for binary files (just like `gzip` is a standard tool for compressing text files). It comes preinstalled on macOS and _most_ Linuxes (at least the ones used for desktop and server use, though not always in containers). Double clicking an `xz` file will automatically unpack it - same as with `zip` or `tar`. ## Size Comparison Using default settings. | | N/A | xz | bzip2 | gzip | | ---------- | ------------ | ---------------- |---------------- | ---------------- | | macOS | 39 MB | 9.8 MB (**75%**) | 14 MB (64%) | 14 MB (64%) | | Linux x86 | 395 MB | 162 MB (**59%**) | 236 MB (40%) | 249 MB (37%) | | Linux ARM | 380 MB | 161 MB (**58%**) | 235 MB (38%) | 249 MB (34%) | I put bzip2 on there for comparison, even though it never really gained traction compared to `xz` and `gzip`, it does come installed by default on most systems. ## Why Compress at All? I realize that among the gigabytes of downloads a couple hundred MB is in the noise, but I'd like to make this available via Webi (<https://webinstall.dev>) and it's very "on-brand" to have things install very quickly with minimal overhead. ## `tar` too? Additionally, if it were packaged as `.tar.xz`, the `execute` bit could be preserved, which would make it just a little bit easier to install (no need to `chmod a+x ./ollama-*`). note for the future: Windows 10 and 11 also have BSD tar.exe preinstalled (for about a decade now) which works for both `.tar` and `.zip` files, and works faster and more reliably that the native PowerShell alternative. ## Compatibility with Automated Tools This could also be in addition to the current release files rather than as a pure replacement so that other Webi-like tools that rely on the current release system aren't broken.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/852/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/852/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/572
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/572/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/572/comments
https://api.github.com/repos/ollama/ollama/issues/572/events
https://github.com/ollama/ollama/issues/572
1,908,543,284
I_kwDOJ0Z1Ps5xwg80
572
How to prevent model answer runaway?
{ "login": "FairyTail2000", "id": 22645621, "node_id": "MDQ6VXNlcjIyNjQ1NjIx", "avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FairyTail2000", "html_url": "https://github.com/FairyTail2000", "followers_url": "https://api.github.com/users/FairyTail2000/followers", "following_url": "https://api.github.com/users/FairyTail2000/following{/other_user}", "gists_url": "https://api.github.com/users/FairyTail2000/gists{/gist_id}", "starred_url": "https://api.github.com/users/FairyTail2000/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/FairyTail2000/subscriptions", "organizations_url": "https://api.github.com/users/FairyTail2000/orgs", "repos_url": "https://api.github.com/users/FairyTail2000/repos", "events_url": "https://api.github.com/users/FairyTail2000/events{/privacy}", "received_events_url": "https://api.github.com/users/FairyTail2000/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
9
2023-09-22T09:30:55
2023-11-09T21:16:28
2023-11-09T21:16:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I've encountered the following issue with some models: The model just keeps answering, providing question in the training format and answering them. This just keeps on going or it stops after one or two extra answers. Example: medllama2:7b-q5_0 docker gpu Question from me (fictional): > A male 23 year old person, complains that when he goes to bed, he can't breath through his nose properly. Mostly during Autumn, Winter and Spring. He says it's most of the time one nostril but sometimes 2. Using Spray helps for a while until it has to be used again. Masks FFP2 and OP Masks help too Answer from the model: > The symptoms you described are typical of hay fever, also known as allergic rhinitis, which is an inflammatory response to airborne particles like pollen and dust mites that occur most often during spring and fall. It can cause nasal congestion, difficulty breathing through the nose and sinuses, sneezing, runny nose, and eye irritation or itchiness. Given your symptoms, we should consider this a possibility. What do you think? Would you like to be referred to an allergist for further testing? Answer: Hay fever (allergic rhinitis) is suspected based on the symptoms. The use of spray and masks can help alleviate symptoms but won't treat the underlying cause. Referral to an allergist would be beneficial in identifying potential allergens and developing a treatment plan specific to your needs. As seen above the model created it's own question from the training set and answered it itself. This can happen mutiple tiimes, mostly in the prompt format This isn't just medllama but also happens with orca mini
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/572/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/572/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2658
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2658/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2658/comments
https://api.github.com/repos/ollama/ollama/issues/2658/events
https://github.com/ollama/ollama/issues/2658
2,147,927,324
I_kwDOJ0Z1Ps6ABsUc
2,658
Ollama should clear temp files on exit.
{ "login": "amnweb", "id": 16545063, "node_id": "MDQ6VXNlcjE2NTQ1MDYz", "avatar_url": "https://avatars.githubusercontent.com/u/16545063?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amnweb", "html_url": "https://github.com/amnweb", "followers_url": "https://api.github.com/users/amnweb/followers", "following_url": "https://api.github.com/users/amnweb/following{/other_user}", "gists_url": "https://api.github.com/users/amnweb/gists{/gist_id}", "starred_url": "https://api.github.com/users/amnweb/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amnweb/subscriptions", "organizations_url": "https://api.github.com/users/amnweb/orgs", "repos_url": "https://api.github.com/users/amnweb/repos", "events_url": "https://api.github.com/users/amnweb/events{/privacy}", "received_events_url": "https://api.github.com/users/amnweb/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-02-22T00:12:32
2024-03-20T15:28:06
2024-03-20T15:28:06
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Found that upon exiting, Ollama does not delete temporary files, but upon starting, Ollama creates new identical files again. in temp folder ''..AppData\Local\Temp" OS: Windows 11 ![image](https://github.com/ollama/ollama/assets/16545063/8831fff1-d684-4217-bc39-a6aaac5624e9)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2658/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2658/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8304
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8304/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8304/comments
https://api.github.com/repos/ollama/ollama/issues/8304/events
https://github.com/ollama/ollama/issues/8304
2,768,646,500
I_kwDOJ0Z1Ps6lBjFk
8,304
Unable to bind 0.0.0.0 port to all interfaces, except localhost
{ "login": "drigio", "id": 22986282, "node_id": "MDQ6VXNlcjIyOTg2Mjgy", "avatar_url": "https://avatars.githubusercontent.com/u/22986282?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drigio", "html_url": "https://github.com/drigio", "followers_url": "https://api.github.com/users/drigio/followers", "following_url": "https://api.github.com/users/drigio/following{/other_user}", "gists_url": "https://api.github.com/users/drigio/gists{/gist_id}", "starred_url": "https://api.github.com/users/drigio/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drigio/subscriptions", "organizations_url": "https://api.github.com/users/drigio/orgs", "repos_url": "https://api.github.com/users/drigio/repos", "events_url": "https://api.github.com/users/drigio/events{/privacy}", "received_events_url": "https://api.github.com/users/drigio/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
6
2025-01-04T08:36:06
2025-01-24T09:46:59
2025-01-24T09:46:59
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm trying to `ollama serve` by setting env variable `OLLAMA_HOST=0.0.0.0` on windows. However I am only able to access it via localhost:11434. Unable to access it via my host ip, nor via 127.0.0.1. There is no proxy enabled, no vpn. I have disabled firewall in my windows machine. ollama serve logs ``` 2025/01/04 00:19:16 routes.go:1259: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\<REDACTED>\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-01-04T00:19:16.089-08:00 level=INFO source=images.go:757 msg="total blobs: 9" time=2025-01-04T00:19:16.090-08:00 level=INFO source=images.go:764 msg="total unused blobs removed: 0" time=2025-01-04T00:19:16.091-08:00 level=INFO source=routes.go:1310 msg="Listening on [::]:11434 (version 0.5.4)" time=2025-01-04T00:19:16.091-08:00 level=INFO source=routes.go:1339 msg="Dynamic LLM libraries" runners="[cuda_v11_avx cuda_v12_avx rocm_avx cpu cpu_avx cpu_avx2]" time=2025-01-04T00:19:16.091-08:00 level=INFO source=gpu.go:226 msg="looking for compatible GPUs" time=2025-01-04T00:19:16.091-08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1 time=2025-01-04T00:19:16.092-08:00 level=INFO source=gpu_windows.go:183 msg="efficiency cores detected" maxEfficiencyClass=1 time=2025-01-04T00:19:16.092-08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=10 efficiency=4 threads=16 ``` ollama list logs ``` ollama list Error: Head "http://0.0.0.0:11434/": read tcp 127.0.0.1:65461->127.0.0.1:11434: wsarecv: An existing connection was forcibly closed by the remote host. $env:OLLAMA_HOST = "localhost"; ollama list NAME ID SIZE MODIFIED nomic-embed-text:latest 0a109f422b47 274 MB About an hour ago llama3.1:8b-instruct-q8_0 b158ded76fa0 8.5 GB 6 hours ago $env:OLLAMA_HOST = "HOST_IP"; ollama list Error: Head "http://HOST_IP:11434/": read tcp HOST_IP:65470->HOST_IP:11434: wsarecv: An existing connection was forcibly closed by the remote host. ``` Can someone please help on this? I am not running WSL. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.4
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8304/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8304/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7425
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7425/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7425/comments
https://api.github.com/repos/ollama/ollama/issues/7425/events
https://github.com/ollama/ollama/issues/7425
2,624,721,955
I_kwDOJ0Z1Ps6cchQj
7,425
Model loaded to RAM and not VRAM
{ "login": "johanesalxd", "id": 17249308, "node_id": "MDQ6VXNlcjE3MjQ5MzA4", "avatar_url": "https://avatars.githubusercontent.com/u/17249308?v=4", "gravatar_id": "", "url": "https://api.github.com/users/johanesalxd", "html_url": "https://github.com/johanesalxd", "followers_url": "https://api.github.com/users/johanesalxd/followers", "following_url": "https://api.github.com/users/johanesalxd/following{/other_user}", "gists_url": "https://api.github.com/users/johanesalxd/gists{/gist_id}", "starred_url": "https://api.github.com/users/johanesalxd/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/johanesalxd/subscriptions", "organizations_url": "https://api.github.com/users/johanesalxd/orgs", "repos_url": "https://api.github.com/users/johanesalxd/repos", "events_url": "https://api.github.com/users/johanesalxd/events{/privacy}", "received_events_url": "https://api.github.com/users/johanesalxd/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" }, { "id": 6433346500, "node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA", "url": "https://api.github.com/repos/ollama/ollama/labels/amd", "name": "amd", "color": "000000", "default": false, "description": "Issues relating to AMD GPUs and ROCm" } ]
closed
false
null
[]
null
1
2024-10-30T16:52:20
2024-10-30T16:55:15
2024-10-30T16:54:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ![image](https://github.com/user-attachments/assets/0b0cad75-7e33-4b6b-8f76-8dc5f9f95b7c) [server.log](https://github.com/user-attachments/files/17576314/server.log) [lmstudio-issues](https://discord.com/channels/1110598183144399058/1294935199892639834) [complete-specs](https://pcpartpicker.com/b/yDZZxr) Model loaded to the RAM and not VRAM (7900 XTX), hence the performance is slow. I'm using Adrenalin 24.10.1 and having the same issue on other app as well (LM Studio) but I'm able to _downgrade it's runtime_ so its running properly again. ![image](https://github.com/user-attachments/assets/0d8985f2-0535-4f4c-82da-0d0b7a961756) ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version 0.3.14
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7425/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7425/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7136
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7136/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7136/comments
https://api.github.com/repos/ollama/ollama/issues/7136/events
https://github.com/ollama/ollama/issues/7136
2,573,570,199
I_kwDOJ0Z1Ps6ZZZCX
7,136
Tool binding prevent streaming
{ "login": "pyautoml", "id": 147545771, "node_id": "U_kgDOCMteqw", "avatar_url": "https://avatars.githubusercontent.com/u/147545771?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pyautoml", "html_url": "https://github.com/pyautoml", "followers_url": "https://api.github.com/users/pyautoml/followers", "following_url": "https://api.github.com/users/pyautoml/following{/other_user}", "gists_url": "https://api.github.com/users/pyautoml/gists{/gist_id}", "starred_url": "https://api.github.com/users/pyautoml/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pyautoml/subscriptions", "organizations_url": "https://api.github.com/users/pyautoml/orgs", "repos_url": "https://api.github.com/users/pyautoml/repos", "events_url": "https://api.github.com/users/pyautoml/events{/privacy}", "received_events_url": "https://api.github.com/users/pyautoml/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-10-08T15:39:32
2024-10-16T01:41:11
2024-10-16T01:41:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi there. I discovered recently that onces a model is delivered tool via bind_tools is stops recognizing streeaming, and each time it waits for the data to be fully loaded. This snippet below works perfectly fine until tools are added: print("AI:", end=" ", flush=True) for chunk in self.model.stream([("human", user_input)], callbacks=[streaming_handler]): if chunk.content: print(chunk.content, end='', flush=True) It happens regardless of the model type. ### OS Windows ### GPU _No response_ ### CPU Intel ### Ollama version 0.3.11
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7136/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/7136/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8118
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8118/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8118/comments
https://api.github.com/repos/ollama/ollama/issues/8118/events
https://github.com/ollama/ollama/pull/8118
2,742,578,632
PR_kwDOJ0Z1Ps6FXfka
8,118
Change the update policy for OLLAMA_ORIGINS to Override
{ "login": "yaoxi-std", "id": 57950207, "node_id": "MDQ6VXNlcjU3OTUwMjA3", "avatar_url": "https://avatars.githubusercontent.com/u/57950207?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yaoxi-std", "html_url": "https://github.com/yaoxi-std", "followers_url": "https://api.github.com/users/yaoxi-std/followers", "following_url": "https://api.github.com/users/yaoxi-std/following{/other_user}", "gists_url": "https://api.github.com/users/yaoxi-std/gists{/gist_id}", "starred_url": "https://api.github.com/users/yaoxi-std/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yaoxi-std/subscriptions", "organizations_url": "https://api.github.com/users/yaoxi-std/orgs", "repos_url": "https://api.github.com/users/yaoxi-std/repos", "events_url": "https://api.github.com/users/yaoxi-std/events{/privacy}", "received_events_url": "https://api.github.com/users/yaoxi-std/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-12-16T14:43:16
2024-12-23T15:41:52
2024-12-23T15:41:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8118", "html_url": "https://github.com/ollama/ollama/pull/8118", "diff_url": "https://github.com/ollama/ollama/pull/8118.diff", "patch_url": "https://github.com/ollama/ollama/pull/8118.patch", "merged_at": null }
So that I will be able to remove some of the default origins.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8118/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8118/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7557
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7557/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7557/comments
https://api.github.com/repos/ollama/ollama/issues/7557/events
https://github.com/ollama/ollama/issues/7557
2,641,000,580
I_kwDOJ0Z1Ps6daniE
7,557
我的文件明明只有12GB,为什么加载的时候,却显示需要22GB显存?
{ "login": "fg2501", "id": 164639270, "node_id": "U_kgDOCdAyJg", "avatar_url": "https://avatars.githubusercontent.com/u/164639270?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fg2501", "html_url": "https://github.com/fg2501", "followers_url": "https://api.github.com/users/fg2501/followers", "following_url": "https://api.github.com/users/fg2501/following{/other_user}", "gists_url": "https://api.github.com/users/fg2501/gists{/gist_id}", "starred_url": "https://api.github.com/users/fg2501/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fg2501/subscriptions", "organizations_url": "https://api.github.com/users/fg2501/orgs", "repos_url": "https://api.github.com/users/fg2501/repos", "events_url": "https://api.github.com/users/fg2501/events{/privacy}", "received_events_url": "https://api.github.com/users/fg2501/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
4
2024-11-07T13:13:24
2024-11-08T03:21:17
2024-11-07T21:50:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ![3f5ba904-d558-4770-b3e3-ee7783694b2f](https://github.com/user-attachments/assets/1a2ef30a-766c-4f00-873c-0b3468238754) ![1111](https://github.com/user-attachments/assets/03dba386-b3cb-4267-9d02-07b048c3a503) 我用的qwen2.5:14b模型,为什么我加载的时候,显示却是需要22GB显存呢? ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.3.13
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7557/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7557/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4825
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4825/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4825/comments
https://api.github.com/repos/ollama/ollama/issues/4825/events
https://github.com/ollama/ollama/issues/4825
2,334,841,826
I_kwDOJ0Z1Ps6LKtvi
4,825
Adapter doesnt work
{ "login": "namogg", "id": 108812102, "node_id": "U_kgDOBnxXRg", "avatar_url": "https://avatars.githubusercontent.com/u/108812102?v=4", "gravatar_id": "", "url": "https://api.github.com/users/namogg", "html_url": "https://github.com/namogg", "followers_url": "https://api.github.com/users/namogg/followers", "following_url": "https://api.github.com/users/namogg/following{/other_user}", "gists_url": "https://api.github.com/users/namogg/gists{/gist_id}", "starred_url": "https://api.github.com/users/namogg/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/namogg/subscriptions", "organizations_url": "https://api.github.com/users/namogg/orgs", "repos_url": "https://api.github.com/users/namogg/repos", "events_url": "https://api.github.com/users/namogg/events{/privacy}", "received_events_url": "https://api.github.com/users/namogg/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
6
2024-06-05T03:54:33
2024-07-10T19:38:32
2024-07-10T18:30:41
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I have converted my adapter to GGML format and create ollama model sucessfully with this comman ADAPTER downloads/ggml-adapter-model.bin However when testing the model, the adapter have no affect on the model. ### OS Linux ### GPU Nvidia ### CPU Intel, AMD ### Ollama version 0.1.29
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4825/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4825/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1641
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1641/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1641/comments
https://api.github.com/repos/ollama/ollama/issues/1641/events
https://github.com/ollama/ollama/issues/1641
2,051,188,544
I_kwDOJ0Z1Ps56QqdA
1,641
Ollama hangs sometimes if it runs out of VRAM
{ "login": "nick-tonjum", "id": 81722223, "node_id": "MDQ6VXNlcjgxNzIyMjIz", "avatar_url": "https://avatars.githubusercontent.com/u/81722223?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nick-tonjum", "html_url": "https://github.com/nick-tonjum", "followers_url": "https://api.github.com/users/nick-tonjum/followers", "following_url": "https://api.github.com/users/nick-tonjum/following{/other_user}", "gists_url": "https://api.github.com/users/nick-tonjum/gists{/gist_id}", "starred_url": "https://api.github.com/users/nick-tonjum/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nick-tonjum/subscriptions", "organizations_url": "https://api.github.com/users/nick-tonjum/orgs", "repos_url": "https://api.github.com/users/nick-tonjum/repos", "events_url": "https://api.github.com/users/nick-tonjum/events{/privacy}", "received_events_url": "https://api.github.com/users/nick-tonjum/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2023-12-20T20:04:18
2024-01-21T01:54:47
2024-01-21T01:54:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi! I just have been having an issue with models that cause the system to run out of VRAM. It usually does the following: 1. (attempt to run a model via api, for example Llama2 70b) 1. ollama-runner tries to load the model into VRAM 2. ollama-runner runs out of VRAM and the process kills 3. the API hangs indefinitely until it is killed (via systemctl restart or killing the docker container if applicable) I don't know why it has to be restarted to process the next request, would it be possible to have a feature where it detects if it runs out of VRAM or crashes and then returns an error via the API and/or auto restarts? This is something I've been running into recently as I only have 24g of VRAM Much appreciated!
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1641/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1641/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5384
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5384/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5384/comments
https://api.github.com/repos/ollama/ollama/issues/5384/events
https://github.com/ollama/ollama/issues/5384
2,381,814,914
I_kwDOJ0Z1Ps6N95yC
5,384
dolphin-phi3 and dolphin-qwen2
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/followers", "following_url": "https://api.github.com/users/olumolu/following{/other_user}", "gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}", "starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/olumolu/subscriptions", "organizations_url": "https://api.github.com/users/olumolu/orgs", "repos_url": "https://api.github.com/users/olumolu/repos", "events_url": "https://api.github.com/users/olumolu/events{/privacy}", "received_events_url": "https://api.github.com/users/olumolu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
4
2024-06-29T15:25:22
2024-07-03T17:42:14
2024-07-03T17:42:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
dolphin-phi3 and qwen2 dolphin https://huggingface.co/cognitivecomputations can we have this models so people can us them like dolphin-llama3
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5384/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5384/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/44
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/44/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/44/comments
https://api.github.com/repos/ollama/ollama/issues/44/events
https://github.com/ollama/ollama/pull/44
1,792,041,442
PR_kwDOJ0Z1Ps5U1i1u
44
add ollama system tray
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-07-06T18:33:01
2023-07-07T17:51:57
2023-07-06T19:22:55
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/44", "html_url": "https://github.com/ollama/ollama/pull/44", "diff_url": "https://github.com/ollama/ollama/pull/44.diff", "patch_url": "https://github.com/ollama/ollama/pull/44.patch", "merged_at": null }
null
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/44/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/44/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5741
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5741/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5741/comments
https://api.github.com/repos/ollama/ollama/issues/5741/events
https://github.com/ollama/ollama/issues/5741
2,412,692,108
I_kwDOJ0Z1Ps6PzsKM
5,741
ROCm Memory Issues with Long Contexts
{ "login": "ProjectMoon", "id": 183856, "node_id": "MDQ6VXNlcjE4Mzg1Ng==", "avatar_url": "https://avatars.githubusercontent.com/u/183856?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ProjectMoon", "html_url": "https://github.com/ProjectMoon", "followers_url": "https://api.github.com/users/ProjectMoon/followers", "following_url": "https://api.github.com/users/ProjectMoon/following{/other_user}", "gists_url": "https://api.github.com/users/ProjectMoon/gists{/gist_id}", "starred_url": "https://api.github.com/users/ProjectMoon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ProjectMoon/subscriptions", "organizations_url": "https://api.github.com/users/ProjectMoon/orgs", "repos_url": "https://api.github.com/users/ProjectMoon/repos", "events_url": "https://api.github.com/users/ProjectMoon/events{/privacy}", "received_events_url": "https://api.github.com/users/ProjectMoon/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
2
2024-07-17T05:37:09
2024-12-19T11:53:09
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Similar (?) to #1952. I've been noticing that ollama will crash when using long context lengths on ROCm. In particular, the most noticeable thing is that I can continue large conversations with ollama from start, while the model remains loaded in memory. But after coming back later and the model needs to reload, it cannot process the context. Here is message I put in 1952. ------ Would like to prod this issue again, as I am still seeing this with GLM4 at 65k context size. Loads fine without much context, but has issues loading larger contexts. I even set the context size to 8k o_O. Important bits: * It looks like GPU VRAM hits 100% but then can't spill over into memory for larger contexts. `rocm-smi` shows VRAM going 98%... 99%.. 100%, then crash. * Forcing GPU layers down to 15 out of 41 and disabling mmap and setting num_batch to 256 for GLM 4 makes VRAM hover around 35%, with 8k context size. * Leaving mmap disabled and num_batch at 256, and letting it load all 41 GPU layers into memory uses 69% VRAM. * Setting num_ctx to 60,000 will still make it try to load all layers into the GPU, and then it crashes because it runs out of VRAM. * Moving num_gpu down to 30 or even 20 allows it to load more context. But this is only delaying the inevitable. Long enough context will = crash. Shouldn't ollama be calculating that it needs to load less layers into the GPU in this situation? Like I can adjust it manually, but if ollama receives num_ctx that'll make the model crash, shouldn't it start using system RAM instead? ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.2.5
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5741/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5741/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8354
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8354/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8354/comments
https://api.github.com/repos/ollama/ollama/issues/8354/events
https://github.com/ollama/ollama/issues/8354
2,776,575,597
I_kwDOJ0Z1Ps6lfy5t
8,354
Dynamic context size in OpenAI API compatibility.
{ "login": "x0wllaar", "id": 10964379, "node_id": "MDQ6VXNlcjEwOTY0Mzc5", "avatar_url": "https://avatars.githubusercontent.com/u/10964379?v=4", "gravatar_id": "", "url": "https://api.github.com/users/x0wllaar", "html_url": "https://github.com/x0wllaar", "followers_url": "https://api.github.com/users/x0wllaar/followers", "following_url": "https://api.github.com/users/x0wllaar/following{/other_user}", "gists_url": "https://api.github.com/users/x0wllaar/gists{/gist_id}", "starred_url": "https://api.github.com/users/x0wllaar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/x0wllaar/subscriptions", "organizations_url": "https://api.github.com/users/x0wllaar/orgs", "repos_url": "https://api.github.com/users/x0wllaar/repos", "events_url": "https://api.github.com/users/x0wllaar/events{/privacy}", "received_events_url": "https://api.github.com/users/x0wllaar/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2025-01-09T00:08:41
2025-01-16T00:01:18
2025-01-16T00:01:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I noticed that the issue (#5356 ) regarding dynamically setting the context size (num_ctx) in OpenAI API was closed with a note saying it wasn't possible due to limitations of the API. However, I'd like to reopen this discussion as there seems to be a using the extra_body parameter available in the OpenAI API clients. This parameter allows to pass arbitrary data/parameters to the endpoint, and will be useful here. It should be possible to pass additional parameters through the API call with the extra_body option, as shown below: ``` client.chat.completions.create( model="phi4", messages=messages, extra_body={"num_ctx": 16384}, ) ``` See https://github.com/openai/openai-python/blob/main/src/openai/resources/chat/completions.py#L102 for normal and https://github.com/openai/openai-python/blob/main/src/openai/resources/beta/chat/completions.py#L101 for structured completion functions. I would also like to note that other implementations of OpenAI API already use this approach, for example, here's vLLM: https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#extra-parameters I'm willing to work on a pull request if you think this approach is feasible. There already is a similar PR #5357, I'm ready to help getting it to a mergeable state. Thank you so much for all your work on Ollama!
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8354/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8354/timeline
null
duplicate
false
https://api.github.com/repos/ollama/ollama/issues/4529
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4529/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4529/comments
https://api.github.com/repos/ollama/ollama/issues/4529/events
https://github.com/ollama/ollama/issues/4529
2,304,933,652
I_kwDOJ0Z1Ps6JYn8U
4,529
error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2'
{ "login": "Anorid", "id": 139095718, "node_id": "U_kgDOCEpupg", "avatar_url": "https://avatars.githubusercontent.com/u/139095718?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Anorid", "html_url": "https://github.com/Anorid", "followers_url": "https://api.github.com/users/Anorid/followers", "following_url": "https://api.github.com/users/Anorid/following{/other_user}", "gists_url": "https://api.github.com/users/Anorid/gists{/gist_id}", "starred_url": "https://api.github.com/users/Anorid/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Anorid/subscriptions", "organizations_url": "https://api.github.com/users/Anorid/orgs", "repos_url": "https://api.github.com/users/Anorid/repos", "events_url": "https://api.github.com/users/Anorid/events{/privacy}", "received_events_url": "https://api.github.com/users/Anorid/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
22
2024-05-20T02:14:35
2024-07-15T09:32:18
2024-06-24T16:48:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I carefully read the contents of the readme's documentation to try and found that something went wrong time=2024-05-20T10:06:02.688+08:00 level=INFO source=server.go:320 msg="starting llama server" cmd="/tmp/ollama2132883000/runners/cuda_v11/ollama_llama_server --model /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 41 --parallel 1 --port 33525" time=2024-05-20T10:06:02.690+08:00 level=INFO source=sched.go:338 msg="loaded runners" count=1 time=2024-05-20T10:06:02.690+08:00 level=INFO source=server.go:504 msg="waiting for llama runner to start responding" time=2024-05-20T10:06:02.691+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="952d03d" tid="140401842012160" timestamp=1716170762 INFO [main] system info | n_threads=64 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="140401842012160" timestamp=1716170762 total_threads=128 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="127" port="33525" tid="140401842012160" timestamp=1716170762 llama_model_loader: loaded meta data with 21 key-value pairs and 483 tensors from /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.name str = merge5-1 llama_model_loader: - kv 2: qwen2.block_count u32 = 40 llama_model_loader: - kv 3: qwen2.context_length u32 = 32768 llama_model_loader: - kv 4: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 5: qwen2.feed_forward_length u32 = 13696 llama_model_loader: - kv 6: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 7: qwen2.attention.head_count_kv u32 = 40 llama_model_loader: - kv 8: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 9: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 10: general.file_type u32 = 2 llama_model_loader: - kv 11: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 12: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 15: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 16: tokenizer.ggml.eos_token_id u32 = 151643 llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 19: tokenizer.chat_template str = {% for message in messages %}{% if lo... llama_model_loader: - kv 20: general.quantization_version u32 = 2 llama_model_loader: - type f32: 201 tensors llama_model_loader: - type q4_0: 281 tensors llama_model_loader: - type q6_K: 1 tensors time=2024-05-20T10:06:02.944+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server loading model" llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' llama_load_model_from_file: exception loading model terminate called after throwing an instance of 'std::runtime_error' what(): error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' time=2024-05-20T10:06:03.285+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error" time=2024-05-20T10:06:03.535+08:00 level=ERROR source=sched.go:344 msg="error loading llama server" error="llama runner process has terminated: signal: aborted (core dumped) " [GIN] 2024/05/20 - 10:06:03 | 500 | 2.178464527s | 127.0.0.1 | POST "/api/chat" time=2024-05-20T10:06:07.831+08:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=41 memory.available="47.3 GiB" memory.required.full="9.7 GiB" memory.required.partial="9.7 GiB" memory.required.kv="1.6 GiB" memory.weights.total="7.2 GiB" memory.weights.repeating="6.6 GiB" memory.weights.nonrepeating="609.1 MiB" memory.graph.full="307.0 MiB" memory.graph.partial="916.1 MiB" time=2024-05-20T10:06:07.832+08:00 level=INFO source=memory.go:133 msg="offload to gpu" layers.requested=-1 layers.real=41 memory.available="47.3 GiB" memory.required.full="9.7 GiB" memory.required.partial="9.7 GiB" memory.required.kv="1.6 GiB" memory.weights.total="7.2 GiB" memory.weights.repeating="6.6 GiB" memory.weights.nonrepeating="609.1 MiB" memory.graph.full="307.0 MiB" memory.graph.partial="916.1 MiB" time=2024-05-20T10:06:07.832+08:00 level=INFO source=server.go:320 msg="starting llama server" cmd="/tmp/ollama2132883000/runners/cuda_v11/ollama_llama_server --model /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 41 --parallel 1 --port 43339" time=2024-05-20T10:06:07.833+08:00 level=INFO source=sched.go:338 msg="loaded runners" count=1 time=2024-05-20T10:06:07.833+08:00 level=INFO source=server.go:504 msg="waiting for llama runner to start responding" time=2024-05-20T10:06:07.833+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="952d03d" tid="140283378036736" timestamp=1716170767 INFO [main] system info | n_threads=64 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="140283378036736" timestamp=1716170767 total_threads=128 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="127" port="43339" tid="140283378036736" timestamp=1716170767 llama_model_loader: loaded meta data with 21 key-value pairs and 483 tensors from /root/autodl-tmp/models/blobs/sha256-1c751709783923dab2b876d5c5c2ca36d4e205cfef7d88988df45752cb91f245 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.name str = merge5-1 llama_model_loader: - kv 2: qwen2.block_count u32 = 40 llama_model_loader: - kv 3: qwen2.context_length u32 = 32768 llama_model_loader: - kv 4: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 5: qwen2.feed_forward_length u32 = 13696 llama_model_loader: - kv 6: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 7: qwen2.attention.head_count_kv u32 = 40 llama_model_loader: - kv 8: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 9: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 10: general.file_type u32 = 2 llama_model_loader: - kv 11: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 12: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 14: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 15: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 16: tokenizer.ggml.eos_token_id u32 = 151643 llama_model_loader: - kv 17: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 19: tokenizer.chat_template str = {% for message in messages %}{% if lo... llama_model_loader: - kv 20: general.quantization_version u32 = 2 llama_model_loader: - type f32: 201 tensors llama_model_loader: - type q4_0: 281 tensors llama_model_loader: - type q6_K: 1 tensors time=2024-05-20T10:06:08.085+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server loading model" llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' llama_load_model_from_file: exception loading model terminate called after throwing an instance of 'std::runtime_error' what(): error loading model vocabulary: unknown pre-tokenizer type: 'qwen2' time=2024-05-20T10:06:08.437+08:00 level=INFO source=server.go:540 msg="waiting for server to become available" status="llm server error" time=2024-05-20T10:06:08.656+08:00 level=WARN source=sched.go:512 msg="gpu VRAM usage didn't recover within timeout" seconds=5.120574757 time=2024-05-20T10:06:08.688+08:00 level=ERROR source=sched.go:344 msg="error loading llama server" error="llama runner process has terminated: signal: aborted (core dumped) " I look at the 4b to 72b of qwen1.5 provided, so this should be provided by the tokenizer as well ### OS Linux ### GPU Nvidia ### CPU Other ### Ollama version client version is 0.1.38
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4529/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4529/timeline
null
completed
false