url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/5915
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5915/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5915/comments
https://api.github.com/repos/ollama/ollama/issues/5915/events
https://github.com/ollama/ollama/pull/5915
2,427,724,333
PR_kwDOJ0Z1Ps52WWxZ
5,915
added tools when `stream=true`
{ "login": "vertrue", "id": 30557724, "node_id": "MDQ6VXNlcjMwNTU3NzI0", "avatar_url": "https://avatars.githubusercontent.com/u/30557724?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vertrue", "html_url": "https://github.com/vertrue", "followers_url": "https://api.github.com/users/vertrue/followers", "following_url": "https://api.github.com/users/vertrue/following{/other_user}", "gists_url": "https://api.github.com/users/vertrue/gists{/gist_id}", "starred_url": "https://api.github.com/users/vertrue/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vertrue/subscriptions", "organizations_url": "https://api.github.com/users/vertrue/orgs", "repos_url": "https://api.github.com/users/vertrue/repos", "events_url": "https://api.github.com/users/vertrue/events{/privacy}", "received_events_url": "https://api.github.com/users/vertrue/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-07-24T14:18:44
2024-07-30T18:17:25
2024-07-30T18:17:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5915", "html_url": "https://github.com/ollama/ollama/pull/5915", "diff_url": "https://github.com/ollama/ollama/pull/5915.diff", "patch_url": "https://github.com/ollama/ollama/pull/5915.patch", "merged_at": null }
fixes #5796 ``` curl --location '127.0.0.1:11434/v1/chat/completions' \ --header 'Content-Type: application/json' \ --data '{ "messages": [ { "content": "You are a helpful AI assistant that can use tools.", "role": "system" }, { "content": "What is function_1(10, 11)? use provided tools", "role": "user" } ], "model": "llama3-groq-tool-use:8b", "logprobs": false, "n": 1, "stream": true, "temperature": 0, "tools": [ { "type": "function", "function": { "name": "function_1", "description": "uses function function_1 for arguments a and b.", "parameters": { "type": "object", "properties": { "a": { "type": "integer" }, "b": { "type": "integer" } }, "required": [ "a", "b" ] } } } ] }' ``` chunks: ``` data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\u003ctool_call\u003e"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\n"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"{\""},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"id"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\":"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" "},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"0"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":","},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" \""},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"name"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\":"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" \""},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"function"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"_"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"1"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\","},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" \""},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"arguments"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\":"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" {\""},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"a"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830176,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\":"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" "},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"10"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":","},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" \""},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"b"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\":"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":" "},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"11"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"}}\n"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"\u003c/tool_call\u003e"},"finish_reason":null}]} data: {"id":"chatcmpl-20","object":"chat.completion.chunk","created":1721830177,"model":"llama3-groq-tool-use:8b","system_fingerprint":"fp_ollama","choices":[{"index":0,"delta":{"role":"assistant","content":"","tool_calls":[{"id":"call_ejieksiz","type":"function","function":{"name":"function_1","arguments":"{\"a\":10,\"b\":11}"}}]},"finish_reason":"tool_calls"}]} data: [DONE] ``` last chunk: ``` { "id": "chatcmpl-20", "object": "chat.completion.chunk", "created": 1721830177, "model": "llama3-groq-tool-use:8b", "system_fingerprint": "fp_ollama", "choices": [ { "index": 0, "delta": { "role": "assistant", "content": "", "tool_calls": [ { "id": "call_ejieksiz", "type": "function", "function": { "name": "function_1", "arguments": "{\"a\":10,\"b\":11}" } } ] }, "finish_reason": "tool_calls" } ] } ``` also works with `/api/chat` also added `"finish_reason": "tool_calls"` for completions
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5915/reactions", "total_count": 18, "+1": 9, "-1": 0, "laugh": 0, "hooray": 5, "confused": 0, "heart": 0, "rocket": 3, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/5915/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7435
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7435/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7435/comments
https://api.github.com/repos/ollama/ollama/issues/7435/events
https://github.com/ollama/ollama/issues/7435
2,625,700,359
I_kwDOJ0Z1Ps6cgQIH
7,435
加载训练生成的gguf,/v1/chat/completions和/api/generate接口返回的结果不一样
{ "login": "czhcc", "id": 4754730, "node_id": "MDQ6VXNlcjQ3NTQ3MzA=", "avatar_url": "https://avatars.githubusercontent.com/u/4754730?v=4", "gravatar_id": "", "url": "https://api.github.com/users/czhcc", "html_url": "https://github.com/czhcc", "followers_url": "https://api.github.com/users/czhcc/followers", "following_url": "https://api.github.com/users/czhcc/following{/other_user}", "gists_url": "https://api.github.com/users/czhcc/gists{/gist_id}", "starred_url": "https://api.github.com/users/czhcc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/czhcc/subscriptions", "organizations_url": "https://api.github.com/users/czhcc/orgs", "repos_url": "https://api.github.com/users/czhcc/repos", "events_url": "https://api.github.com/users/czhcc/events{/privacy}", "received_events_url": "https://api.github.com/users/czhcc/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-10-31T01:08:25
2024-11-17T14:17:43
2024-11-17T14:17:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 训练的基础模型是qwen2.5-7b。 对于同样的内容, /v1/chat/completions和/api/generate接口返回不同的结果。 /api/generate返回是正确的训练要求的结果。 ### OS Linux, Docker ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.14
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7435/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7435/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8260
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8260/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8260/comments
https://api.github.com/repos/ollama/ollama/issues/8260/events
https://github.com/ollama/ollama/issues/8260
2,761,465,899
I_kwDOJ0Z1Ps6kmKAr
8,260
For CPU systems, getting files which are being used for GPU
{ "login": "Abubakkar13", "id": 45032674, "node_id": "MDQ6VXNlcjQ1MDMyNjc0", "avatar_url": "https://avatars.githubusercontent.com/u/45032674?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Abubakkar13", "html_url": "https://github.com/Abubakkar13", "followers_url": "https://api.github.com/users/Abubakkar13/followers", "following_url": "https://api.github.com/users/Abubakkar13/following{/other_user}", "gists_url": "https://api.github.com/users/Abubakkar13/gists{/gist_id}", "starred_url": "https://api.github.com/users/Abubakkar13/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Abubakkar13/subscriptions", "organizations_url": "https://api.github.com/users/Abubakkar13/orgs", "repos_url": "https://api.github.com/users/Abubakkar13/repos", "events_url": "https://api.github.com/users/Abubakkar13/events{/privacy}", "received_events_url": "https://api.github.com/users/Abubakkar13/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-12-28T04:56:52
2024-12-29T03:14:47
2024-12-29T03:14:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hey, Happy holidays 😍!!! I just came across with Ollama version update, then saw some files are being downloaded, I noticed that few dll libraries are being download, So i just cross checked the folder there are 7 files(Specifically starting with **cuda*** or **cublas*** ) and in runners folder as well, have downloaded. But these are for GPU acceleration, since i am with CPU only, is their a way to ignore these files or is their a way to do a system check just before installing the Ollama so can ignore while installing itself, so it can reduce 2.8GB of space for CPU users? Folder is: 1. Taking around 880M: C:\Users\[USER}\AppData\Local\Programs\Ollama\lib\ollama ![image](https://github.com/user-attachments/assets/3982eb88-d4c9-4283-ab76-099c7873cb94) 2. Taking around 2GB: C:\Users\[USER]\AppData\Local\Programs\Ollama\lib\ollama\runners ![image](https://github.com/user-attachments/assets/37f259f6-4c6f-4d25-b245-a815eba34ace) Thank you From GPU Poor 😅 ### OS Windows ### GPU _No response_ ### CPU Intel ### Ollama version 0.5.4
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8260/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8260/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/382
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/382/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/382/comments
https://api.github.com/repos/ollama/ollama/issues/382/events
https://github.com/ollama/ollama/pull/382
1,857,140,673
PR_kwDOJ0Z1Ps5YRY9L
382
Closes #371
{ "login": "jesjess243", "id": 123990927, "node_id": "U_kgDOB2Pzjw", "avatar_url": "https://avatars.githubusercontent.com/u/123990927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jesjess243", "html_url": "https://github.com/jesjess243", "followers_url": "https://api.github.com/users/jesjess243/followers", "following_url": "https://api.github.com/users/jesjess243/following{/other_user}", "gists_url": "https://api.github.com/users/jesjess243/gists{/gist_id}", "starred_url": "https://api.github.com/users/jesjess243/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jesjess243/subscriptions", "organizations_url": "https://api.github.com/users/jesjess243/orgs", "repos_url": "https://api.github.com/users/jesjess243/repos", "events_url": "https://api.github.com/users/jesjess243/events{/privacy}", "received_events_url": "https://api.github.com/users/jesjess243/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-08-18T18:43:29
2023-08-22T19:21:00
2023-08-22T19:21:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/382", "html_url": "https://github.com/ollama/ollama/pull/382", "diff_url": "https://github.com/ollama/ollama/pull/382.diff", "patch_url": "https://github.com/ollama/ollama/pull/382.patch", "merged_at": null }
Hello, ModelPath.ParseModelPath(name) now trims https:// from name if it is present. Keeping in the scope of the issue, no alternative protocols are accepted and a path with no protocol defaults to https as before. My first approach would've allowed garbage text in the protocol, but this lacks that side-effect.
{ "login": "jesjess243", "id": 123990927, "node_id": "U_kgDOB2Pzjw", "avatar_url": "https://avatars.githubusercontent.com/u/123990927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jesjess243", "html_url": "https://github.com/jesjess243", "followers_url": "https://api.github.com/users/jesjess243/followers", "following_url": "https://api.github.com/users/jesjess243/following{/other_user}", "gists_url": "https://api.github.com/users/jesjess243/gists{/gist_id}", "starred_url": "https://api.github.com/users/jesjess243/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jesjess243/subscriptions", "organizations_url": "https://api.github.com/users/jesjess243/orgs", "repos_url": "https://api.github.com/users/jesjess243/repos", "events_url": "https://api.github.com/users/jesjess243/events{/privacy}", "received_events_url": "https://api.github.com/users/jesjess243/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/382/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/382/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6473
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6473/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6473/comments
https://api.github.com/repos/ollama/ollama/issues/6473/events
https://github.com/ollama/ollama/issues/6473
2,482,749,930
I_kwDOJ0Z1Ps6T-8Hq
6,473
OpenAI Structured Output Compatability
{ "login": "jd-solanki", "id": 47495003, "node_id": "MDQ6VXNlcjQ3NDk1MDAz", "avatar_url": "https://avatars.githubusercontent.com/u/47495003?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jd-solanki", "html_url": "https://github.com/jd-solanki", "followers_url": "https://api.github.com/users/jd-solanki/followers", "following_url": "https://api.github.com/users/jd-solanki/following{/other_user}", "gists_url": "https://api.github.com/users/jd-solanki/gists{/gist_id}", "starred_url": "https://api.github.com/users/jd-solanki/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jd-solanki/subscriptions", "organizations_url": "https://api.github.com/users/jd-solanki/orgs", "repos_url": "https://api.github.com/users/jd-solanki/repos", "events_url": "https://api.github.com/users/jd-solanki/events{/privacy}", "received_events_url": "https://api.github.com/users/jd-solanki/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
9
2024-08-23T09:42:33
2024-12-05T00:45:49
2024-12-05T00:45:49
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi 👋🏻 Loving ollama always ❤️ I'm eager to use newly released structured output using ollama but it looks like ollama doesn't have compatibility yet so I can just put base_url and I'll get response from my local LLM. Also, I would like will it support [streaming](https://python.useinstructor.com/concepts/partial/) like instructor? Thanks.
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6473/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6473/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4292
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4292/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4292/comments
https://api.github.com/repos/ollama/ollama/issues/4292/events
https://github.com/ollama/ollama/pull/4292
2,288,040,474
PR_kwDOJ0Z1Ps5vAnA9
4,292
remove last bits of ParseModelPath
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
1
2024-05-09T16:25:46
2024-08-29T23:26:19
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4292", "html_url": "https://github.com/ollama/ollama/pull/4292", "diff_url": "https://github.com/ollama/ollama/pull/4292.diff", "patch_url": "https://github.com/ollama/ollama/pull/4292.patch", "merged_at": null }
follow up to #4291
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4292/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4292/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5455
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5455/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5455/comments
https://api.github.com/repos/ollama/ollama/issues/5455/events
https://github.com/ollama/ollama/issues/5455
2,388,109,167
I_kwDOJ0Z1Ps6OV6dv
5,455
ollama does not work on ALL GPU automatically
{ "login": "HeroSong666", "id": 142960235, "node_id": "U_kgDOCIVmaw", "avatar_url": "https://avatars.githubusercontent.com/u/142960235?v=4", "gravatar_id": "", "url": "https://api.github.com/users/HeroSong666", "html_url": "https://github.com/HeroSong666", "followers_url": "https://api.github.com/users/HeroSong666/followers", "following_url": "https://api.github.com/users/HeroSong666/following{/other_user}", "gists_url": "https://api.github.com/users/HeroSong666/gists{/gist_id}", "starred_url": "https://api.github.com/users/HeroSong666/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HeroSong666/subscriptions", "organizations_url": "https://api.github.com/users/HeroSong666/orgs", "repos_url": "https://api.github.com/users/HeroSong666/repos", "events_url": "https://api.github.com/users/HeroSong666/events{/privacy}", "received_events_url": "https://api.github.com/users/HeroSong666/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
5
2024-07-03T09:12:14
2024-11-08T08:27:58
2024-08-01T23:51:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? when I use the ollama:0.1.38, I use the following command to start: ` docker run -d --gpus=all -v /root/ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama ` ollama will automatically use all 4 gpu cards for inference. when I upgrate to ollama:0.1.48, I use the same command, but it only use 1 gpu for inference: ![1](https://github.com/ollama/ollama/assets/142960235/b7a0d651-067e-4670-9a58-c02c5076f046) I remember the running 'Processes' is not '...unners/cuda_v11/ollama_llama_server', should be "ollama/ollama" or something else before. Why this happens? Also, when I use ` docker run -d --gpus=all -v /root/ollama:/root/.ollama -p 11434:11434 -e OLLAMA_SCHED_SPREAD=1 --name ollama ollama/ollama:0.1.48 ` to force it use all 4 gpus for inference, I noticed that the combined utilization of the four GPU cards will not reach 100%, let alone 400%. Here is the gpu usage I monitor. [gpu_usage.csv](https://github.com/user-attachments/files/16081349/gpu_usage.csv) I think ollama does not make full use of GPU resources. Why is this? ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.48
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5455/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5455/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6939
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6939/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6939/comments
https://api.github.com/repos/ollama/ollama/issues/6939/events
https://github.com/ollama/ollama/pull/6939
2,546,109,699
PR_kwDOJ0Z1Ps58jqoL
6,939
docs: update llamaindex links
{ "login": "himself65", "id": 14026360, "node_id": "MDQ6VXNlcjE0MDI2MzYw", "avatar_url": "https://avatars.githubusercontent.com/u/14026360?v=4", "gravatar_id": "", "url": "https://api.github.com/users/himself65", "html_url": "https://github.com/himself65", "followers_url": "https://api.github.com/users/himself65/followers", "following_url": "https://api.github.com/users/himself65/following{/other_user}", "gists_url": "https://api.github.com/users/himself65/gists{/gist_id}", "starred_url": "https://api.github.com/users/himself65/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/himself65/subscriptions", "organizations_url": "https://api.github.com/users/himself65/orgs", "repos_url": "https://api.github.com/users/himself65/repos", "events_url": "https://api.github.com/users/himself65/events{/privacy}", "received_events_url": "https://api.github.com/users/himself65/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-09-24T18:44:21
2024-09-24T19:15:59
2024-09-24T19:15:43
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6939", "html_url": "https://github.com/ollama/ollama/pull/6939", "diff_url": "https://github.com/ollama/ollama/pull/6939.diff", "patch_url": "https://github.com/ollama/ollama/pull/6939.patch", "merged_at": "2024-09-24T19:15:43" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6939/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6939/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1756
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1756/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1756/comments
https://api.github.com/repos/ollama/ollama/issues/1756/events
https://github.com/ollama/ollama/issues/1756
2,061,766,722
I_kwDOJ0Z1Ps565BBC
1,756
Older CUDA compute capability 3.5 and 3.7 support
{ "login": "orlyandico", "id": 1325420, "node_id": "MDQ6VXNlcjEzMjU0MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orlyandico", "html_url": "https://github.com/orlyandico", "followers_url": "https://api.github.com/users/orlyandico/followers", "following_url": "https://api.github.com/users/orlyandico/following{/other_user}", "gists_url": "https://api.github.com/users/orlyandico/gists{/gist_id}", "starred_url": "https://api.github.com/users/orlyandico/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/orlyandico/subscriptions", "organizations_url": "https://api.github.com/users/orlyandico/orgs", "repos_url": "https://api.github.com/users/orlyandico/repos", "events_url": "https://api.github.com/users/orlyandico/events{/privacy}", "received_events_url": "https://api.github.com/users/orlyandico/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 7700262114, "node_id": "LA_kwDOJ0Z1Ps8AAAAByvis4g", "url": "https://api.github.com/repos/ollama/ollama/labels/build", "name": "build", "color": "006b75", "default": false, "description": "Issues relating to building ollama from source" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
64
2024-01-01T20:49:39
2025-01-27T15:35:54
2024-12-10T17:47:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I recently put together an (old) physical machine with an Nvidia K80, which is only supported up to CUDA 11.4 and Nvidia driver 470. All my previous experiments with Ollama were with more modern GPU's. I found that Ollama doesn't use the GPU at all. I cannot find any documentation on the minimum required CUDA version, and if it is possible to run on older CUDA versions (e.g. Nvidia K80, V100 are still present on cloud, e.g. G2 and P2 on AWS) and there's lots of K80's all over ebay. EDIT: looking through the logs, it appears that the GPU's are being seen: Jan 1 20:22:43 thinkstation-s30 ollama[911]: 2024/01/01 20:22:43 llama.go:300: 24762 MB VRAM available, loading up to 162 GPU layers Jan 1 20:22:43 thinkstation-s30 ollama[911]: 2024/01/01 20:22:43 llama.go:436: starting llama runner Jan 1 20:22:43 thinkstation-s30 ollama[911]: 2024/01/01 20:22:43 llama.go:494: waiting for llama runner to start responding Jan 1 20:22:43 thinkstation-s30 ollama[911]: ggml_init_cublas: GGML_CUDA_FORCE_MMQ: no Jan 1 20:22:43 thinkstation-s30 ollama[911]: ggml_init_cublas: CUDA_USE_TENSOR_CORES: yes Jan 1 20:22:43 thinkstation-s30 ollama[911]: ggml_init_cublas: found 3 CUDA devices: Jan 1 20:22:43 thinkstation-s30 ollama[911]: Device 0: Tesla K80, compute capability 3.7 Jan 1 20:22:43 thinkstation-s30 ollama[911]: Device 1: Tesla K80, compute capability 3.7 Jan 1 20:22:43 thinkstation-s30 ollama[911]: Device 2: NVIDIA GeForce GT 730, compute capability 3.5 and Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: ggml ctx size = 0.11 MiB Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: using CUDA for GPU acceleration Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: mem required = 70.46 MiB Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: offloading 32 repeating layers to GPU Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: offloading non-repeating layers to GPU Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: offloaded 33/33 layers to GPU Jan 1 20:34:20 thinkstation-s30 ollama[911]: llm_load_tensors: VRAM used: 3577.61 MiB but.... Jan 1 20:34:21 thinkstation-s30 ollama[911]: CUDA error 209 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:7801: no kernel image is available for execution on the device Jan 1 20:34:21 thinkstation-s30 ollama[911]: current device: 0 Jan 1 20:34:21 thinkstation-s30 ollama[911]: GGML_ASSERT: /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:7801: !"CUDA error" Jan 1 20:34:22 thinkstation-s30 ollama[911]: 2024/01/01 20:34:22 llama.go:451: 209 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:7801: no kernel image is available for execution on the device Jan 1 20:34:22 thinkstation-s30 ollama[911]: current device: 0 Jan 1 20:34:22 thinkstation-s30 ollama[911]: GGML_ASSERT: /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:7801: !"CUDA error" Jan 1 20:34:22 thinkstation-s30 ollama[911]: 2024/01/01 20:34:22 llama.go:459: error starting llama runner: llama runner process has terminated
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1756/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1756/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8543
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8543/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8543/comments
https://api.github.com/repos/ollama/ollama/issues/8543/events
https://github.com/ollama/ollama/issues/8543
2,805,677,144
I_kwDOJ0Z1Ps6nOzxY
8,543
Ollama does not perform structured output correctly.
{ "login": "aJarOfCode", "id": 103819351, "node_id": "U_kgDOBjAoVw", "avatar_url": "https://avatars.githubusercontent.com/u/103819351?v=4", "gravatar_id": "", "url": "https://api.github.com/users/aJarOfCode", "html_url": "https://github.com/aJarOfCode", "followers_url": "https://api.github.com/users/aJarOfCode/followers", "following_url": "https://api.github.com/users/aJarOfCode/following{/other_user}", "gists_url": "https://api.github.com/users/aJarOfCode/gists{/gist_id}", "starred_url": "https://api.github.com/users/aJarOfCode/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/aJarOfCode/subscriptions", "organizations_url": "https://api.github.com/users/aJarOfCode/orgs", "repos_url": "https://api.github.com/users/aJarOfCode/repos", "events_url": "https://api.github.com/users/aJarOfCode/events{/privacy}", "received_events_url": "https://api.github.com/users/aJarOfCode/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-23T01:19:21
2025-01-27T09:10:56
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? request: `{ "model":"llama3.2", "messages":datas+[ { "role":"user", "content":input_data, "images":[screenshot_base64] } ], "stream":False, "format":{ "type":"object", "reply":{"type":"string"}, "properties":{ "operations":{"type":"array", "instruct":{ "type":"object", "functions":{"type":"object", "function_name":{ "type":"string"}, "parameter":{"type":"array","items":{"type":"string"}} } } } } }, "required":["reply"] }` reply: `{"operations": []}` Ollama did not reply with the required "reply". ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8543/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8543/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7615
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7615/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7615/comments
https://api.github.com/repos/ollama/ollama/issues/7615/events
https://github.com/ollama/ollama/pull/7615
2,648,429,620
PR_kwDOJ0Z1Ps6BeCEy
7,615
Allow Compile on older GPUs - still on CUDA 11.3
{ "login": "langstonmeister", "id": 65471211, "node_id": "MDQ6VXNlcjY1NDcxMjEx", "avatar_url": "https://avatars.githubusercontent.com/u/65471211?v=4", "gravatar_id": "", "url": "https://api.github.com/users/langstonmeister", "html_url": "https://github.com/langstonmeister", "followers_url": "https://api.github.com/users/langstonmeister/followers", "following_url": "https://api.github.com/users/langstonmeister/following{/other_user}", "gists_url": "https://api.github.com/users/langstonmeister/gists{/gist_id}", "starred_url": "https://api.github.com/users/langstonmeister/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/langstonmeister/subscriptions", "organizations_url": "https://api.github.com/users/langstonmeister/orgs", "repos_url": "https://api.github.com/users/langstonmeister/repos", "events_url": "https://api.github.com/users/langstonmeister/events{/privacy}", "received_events_url": "https://api.github.com/users/langstonmeister/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
1
2024-11-11T07:31:03
2024-11-22T17:01:26
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7615", "html_url": "https://github.com/ollama/ollama/pull/7615", "diff_url": "https://github.com/ollama/ollama/pull/7615.diff", "patch_url": "https://github.com/ollama/ollama/pull/7615.patch", "merged_at": null }
My Tesla k40s are still able to run this version of CUDA, but are technically on Compute 3.5. These commits will allow them to compile and run the software, and are working well for me.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7615/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7615/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2331
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2331/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2331/comments
https://api.github.com/repos/ollama/ollama/issues/2331/events
https://github.com/ollama/ollama/pull/2331
2,115,267,354
PR_kwDOJ0Z1Ps5l2Y_S
2,331
Update README.md
{ "login": "AshD", "id": 233016, "node_id": "MDQ6VXNlcjIzMzAxNg==", "avatar_url": "https://avatars.githubusercontent.com/u/233016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AshD", "html_url": "https://github.com/AshD", "followers_url": "https://api.github.com/users/AshD/followers", "following_url": "https://api.github.com/users/AshD/following{/other_user}", "gists_url": "https://api.github.com/users/AshD/gists{/gist_id}", "starred_url": "https://api.github.com/users/AshD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AshD/subscriptions", "organizations_url": "https://api.github.com/users/AshD/orgs", "repos_url": "https://api.github.com/users/AshD/repos", "events_url": "https://api.github.com/users/AshD/events{/privacy}", "received_events_url": "https://api.github.com/users/AshD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-02-02T16:12:29
2024-11-21T09:13:03
2024-11-21T09:13:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2331", "html_url": "https://github.com/ollama/ollama/pull/2331", "diff_url": "https://github.com/ollama/ollama/pull/2331.diff", "patch_url": "https://github.com/ollama/ollama/pull/2331.patch", "merged_at": null }
Adding info on Fusion Quill. Setup info is here https://fusionquill.ai/help-setup-ollama/ Fusion Quill Personal Edition is a Windows app on the Microsoft App Store that connects to multiple AI models with workflows and UX like an Integrated Word processor with AI Chat in a split-pane UI that enables creating documents with ease! Fusion Quill Personal Edition supports AI Writing Buddy with Multiple AIs like Ollama, OpenAI, Mistral, Azure AI, Google Gemini, Bedrock, vLLM, etc. Chat with a Debate Coach, Interview Coach and other assistants. More info at https://FusionQuill.AI
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2331/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2331/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4268
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4268/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4268/comments
https://api.github.com/repos/ollama/ollama/issues/4268/events
https://github.com/ollama/ollama/pull/4268
2,286,626,904
PR_kwDOJ0Z1Ps5u72dc
4,268
Convert directly from llama3
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-05-08T23:14:20
2024-05-21T21:43:38
2024-05-21T21:43:37
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4268", "html_url": "https://github.com/ollama/ollama/pull/4268", "diff_url": "https://github.com/ollama/ollama/pull/4268.diff", "patch_url": "https://github.com/ollama/ollama/pull/4268.patch", "merged_at": "2024-05-21T21:43:37" }
This change allows you to convert directly from a llama3 derived safetensors model into Ollama. It is currently *missing*: * pytorch *almost* works however the embeddings layer size is off by the eos/bos tokens This *will* work with most llama3 derivatives if they are using safetensors including `dolphin-2.9-llama3`, nous research's hermes 2 pro, and nvidia's chatqa.
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4268/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4268/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1651
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1651/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1651/comments
https://api.github.com/repos/ollama/ollama/issues/1651/events
https://github.com/ollama/ollama/issues/1651
2,051,690,293
I_kwDOJ0Z1Ps56Sk81
1,651
GPU not in use?
{ "login": "Y2K350", "id": 140640993, "node_id": "U_kgDOCGIC4Q", "avatar_url": "https://avatars.githubusercontent.com/u/140640993?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Y2K350", "html_url": "https://github.com/Y2K350", "followers_url": "https://api.github.com/users/Y2K350/followers", "following_url": "https://api.github.com/users/Y2K350/following{/other_user}", "gists_url": "https://api.github.com/users/Y2K350/gists{/gist_id}", "starred_url": "https://api.github.com/users/Y2K350/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Y2K350/subscriptions", "organizations_url": "https://api.github.com/users/Y2K350/orgs", "repos_url": "https://api.github.com/users/Y2K350/repos", "events_url": "https://api.github.com/users/Y2K350/events{/privacy}", "received_events_url": "https://api.github.com/users/Y2K350/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
25
2023-12-21T04:51:03
2025-01-03T21:54:51
2024-01-10T15:08:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I am running Ollama which was installed on an arch linux system using "sudo pacman -S ollama" I am using a RTX 4090 with Nvidia's latest drivers. I also installed cuda using "sudo pacman -S cuda" I run the LLM using the command "ollama run dolphin-mixtral:latest" it does not appear to use the GPU based on GPU usage provided by GreenWithEnvy (GWE), but I am unsure how to verify that information. Is there a specific command I need to run to ensure it uses the GPU instead of the CPU? nvidia-smi returns: +---------------------------------------------------------------------------------------+ | NVIDIA-SMI 545.29.06 Driver Version: 545.29.06 CUDA Version: 12.3 | |-----------------------------------------+----------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+======================+======================| | 0 NVIDIA GeForce RTX 4090 Off | 00000000:01:00.0 On | Off | | 0% 42C P8 37W / 450W | 20102MiB / 24564MiB | 11% Default | | | | N/A | +-----------------------------------------+----------------------+----------------------+ +---------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=======================================================================================| | 0 N/A N/A 3261 G /usr/lib/Xorg 632MiB | | 0 N/A N/A 3360 G /usr/bin/gnome-shell 327MiB | | 0 N/A N/A 3778 G /usr/lib/xdg-desktop-portal-gnome 14MiB | | 0 N/A N/A 3973 G gjs 99MiB | | 0 N/A N/A 4337 G /app/lib/librewolf/librewolf 219MiB | | 0 N/A N/A 5088 G ...,WinRetrieveSuggestionsOnlyOnDemand 167MiB | | 0 N/A N/A 67846 G /usr/bin/gnome-text-editor 13MiB | | 0 N/A N/A 69685 C ...p/gguf/build/cuda/bin/ollama-runner 18000MiB | | 0 N/A N/A 101455 G /usr/bin/nautilus 32MiB | | 0 N/A N/A 101815 G /usr/bin/kgx 549MiB | +---------------------------------------------------------------------------------------+
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1651/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1651/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1930
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1930/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1930/comments
https://api.github.com/repos/ollama/ollama/issues/1930/events
https://github.com/ollama/ollama/issues/1930
2,077,220,925
I_kwDOJ0Z1Ps57z-A9
1,930
Support for CogVLM wanted. CogVLM is an alternative for LLaVA
{ "login": "henryclw", "id": 97804910, "node_id": "U_kgDOBdRibg", "avatar_url": "https://avatars.githubusercontent.com/u/97804910?v=4", "gravatar_id": "", "url": "https://api.github.com/users/henryclw", "html_url": "https://github.com/henryclw", "followers_url": "https://api.github.com/users/henryclw/followers", "following_url": "https://api.github.com/users/henryclw/following{/other_user}", "gists_url": "https://api.github.com/users/henryclw/gists{/gist_id}", "starred_url": "https://api.github.com/users/henryclw/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/henryclw/subscriptions", "organizations_url": "https://api.github.com/users/henryclw/orgs", "repos_url": "https://api.github.com/users/henryclw/repos", "events_url": "https://api.github.com/users/henryclw/events{/privacy}", "received_events_url": "https://api.github.com/users/henryclw/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
7
2024-01-11T17:23:22
2024-06-24T18:14:00
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Currently ollama is supporting LLaVA, which is super great. I wonder is there a chance to load other similar models like CogVLM? https://github.com/THUDM/CogVLM
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1930/reactions", "total_count": 6, "+1": 6, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1930/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5525
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5525/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5525/comments
https://api.github.com/repos/ollama/ollama/issues/5525/events
https://github.com/ollama/ollama/issues/5525
2,393,903,563
I_kwDOJ0Z1Ps6OsBHL
5,525
Suggestions
{ "login": "EchoOfMedivhCheats", "id": 174923936, "node_id": "U_kgDOCm0goA", "avatar_url": "https://avatars.githubusercontent.com/u/174923936?v=4", "gravatar_id": "", "url": "https://api.github.com/users/EchoOfMedivhCheats", "html_url": "https://github.com/EchoOfMedivhCheats", "followers_url": "https://api.github.com/users/EchoOfMedivhCheats/followers", "following_url": "https://api.github.com/users/EchoOfMedivhCheats/following{/other_user}", "gists_url": "https://api.github.com/users/EchoOfMedivhCheats/gists{/gist_id}", "starred_url": "https://api.github.com/users/EchoOfMedivhCheats/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EchoOfMedivhCheats/subscriptions", "organizations_url": "https://api.github.com/users/EchoOfMedivhCheats/orgs", "repos_url": "https://api.github.com/users/EchoOfMedivhCheats/repos", "events_url": "https://api.github.com/users/EchoOfMedivhCheats/events{/privacy}", "received_events_url": "https://api.github.com/users/EchoOfMedivhCheats/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-07-07T05:43:01
2024-07-07T05:43:01
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Terminal color/bold codes (needs 3 and 3.%) Custom terminal app (Vulkan for stuff like svgs? Maybe for videos?)) Selinux for English (nouns verbs pronouns) Voice command and output Don’t know that much assembly but I think you can daisy chain stuff I know this is written in go, but a scripting language (if x = chem(h202)) for models haven’t tried to look at them, on the fly compression with etc, some sort of index or master index to cache, instead of printing word by word, you can print the entire thing with a progress bar (or a friendly logo!), and a custom title bar and buttons, and skins
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5525/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5525/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8690
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8690/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8690/comments
https://api.github.com/repos/ollama/ollama/issues/8690/events
https://github.com/ollama/ollama/issues/8690
2,820,660,880
I_kwDOJ0Z1Ps6oH96Q
8,690
Deepseek-671B: Error: timed out waiting for llama runner to start - progress 0.00 on 8x L40S
{ "login": "orlyandico", "id": 1325420, "node_id": "MDQ6VXNlcjEzMjU0MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orlyandico", "html_url": "https://github.com/orlyandico", "followers_url": "https://api.github.com/users/orlyandico/followers", "following_url": "https://api.github.com/users/orlyandico/following{/other_user}", "gists_url": "https://api.github.com/users/orlyandico/gists{/gist_id}", "starred_url": "https://api.github.com/users/orlyandico/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/orlyandico/subscriptions", "organizations_url": "https://api.github.com/users/orlyandico/orgs", "repos_url": "https://api.github.com/users/orlyandico/repos", "events_url": "https://api.github.com/users/orlyandico/events{/privacy}", "received_events_url": "https://api.github.com/users/orlyandico/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-30T12:08:46
2025-01-30T12:12:22
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Ollama (0.5.7) appears to be correctly calculating how many layers to offload to the GPU with default settings. This is on a g6e.48xlarge which has 1.5TB of RAM. ``` Jan 30 11:56:19 ip-172-31-21-180 ollama[3237]: time=2025-01-30T11:56:19.283Z level=INFO source=memory.go:356 msg="offload to cuda" layers.requested=-1 layers.model=62 layers.offload=51 layers.split=7,7,7,6,6,6,6,6 memory.available="[43.9 GiB 43.9 GiB 43.9 GiB 43.9 GiB 43.9 GiB 43.9 GiB 43.9 GiB 43.9 GiB]" memory.gpu_overhead="0 B" memory.required.full="402.1 GiB" memory.required.partial="330.4 GiB" memory.required.kv="9.5 GiB" memory.required.allocations="[41.4 GiB 41.4 GiB 41.4 GiB 40.9 GiB 41.8 GiB 41.8 GiB 40.9 GiB 40.9 GiB]" memory.weights.total="385.0 GiB" memory.weights.repeating="384.3 GiB" memory.weights.nonrepeating="725.0 MiB" memory.graph.full="1019.5 MiB" memory.graph.partial="1019.5 MiB" Jan 30 11:56:19 ip-172-31-21-180 ollama[3237]: time=2025-01-30T11:56:19.284Z level=INFO source=server.go:376 msg="starting llama server" cmd="/usr/local/lib/ollama/runners/cuda_v12_avx/ollama_llama_server runner --model /usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 --ctx-size 2048 --batch-size 512 --n-gpu-layers 51 --threads 96 --parallel 1 --tensor-split 7,7,7,6,6,6,6,6 --port 39933" ... Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA0 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA1 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA2 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA3 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA4 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA5 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA6 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_load_model_from_file: using device CUDA7 (NVIDIA L40S) - 44940 MiB free Jan 30 11:56:20 ip-172-31-21-180 ollama[3237]: llama_model_loader: loaded meta data with 42 key-value pairs and 1025 tensors from /usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 (version GGUF V3 (latest)) ``` However, I never see the GPU VRAM usage climbing (this normally happens on my 2 x P40 setup as the model loads into VRAM) it is stuck at this: ``` Thu Jan 30 12:06:42 2025 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.144.03 Driver Version: 550.144.03 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA L40S On | 00000000:9E:00.0 Off | 0 | | N/A 40C P0 81W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA L40S On | 00000000:A0:00.0 Off | 0 | | N/A 43C P0 87W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 2 NVIDIA L40S On | 00000000:A2:00.0 Off | 0 | | N/A 41C P0 84W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 3 NVIDIA L40S On | 00000000:A4:00.0 Off | 0 | | N/A 40C P0 81W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 4 NVIDIA L40S On | 00000000:C6:00.0 Off | 0 | | N/A 40C P0 79W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 5 NVIDIA L40S On | 00000000:C8:00.0 Off | 0 | | N/A 40C P0 80W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 6 NVIDIA L40S On | 00000000:CA:00.0 Off | 0 | | N/A 40C P0 81W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 7 NVIDIA L40S On | 00000000:CC:00.0 Off | 0 | | N/A 39C P0 81W / 350W | 433MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | 0 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 1 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 2 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 3 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 4 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 5 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 6 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | | 7 N/A N/A 4939 C ...rs/cuda_v12_avx/ollama_llama_server 424MiB | +-----------------------------------------------------------------------------------------+ ``` and at the very end I get this error: ``` Jan 30 12:01:19 ip-172-31-21-180 ollama[3237]: time=2025-01-30T12:01:19.487Z level=ERROR source=sched.go:455 msg="error loading llama server" error="timed out waiting for llama runner to start - progress 0.00 - " Jan 30 12:01:19 ip-172-31-21-180 ollama[3237]: [GIN] 2025/01/30 - 12:01:19 | 500 | 5m4s | 127.0.0.1 | POST "/api/generate" Jan 30 12:01:26 ip-172-31-21-180 ollama[3237]: time=2025-01-30T12:01:26.104Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=6.61651503 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 30 12:01:28 ip-172-31-21-180 ollama[3237]: time=2025-01-30T12:01:28.080Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=8.592545492 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 30 12:01:30 ip-172-31-21-180 ollama[3237]: time=2025-01-30T12:01:30.058Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=10.570809357 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 ``` ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8690/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8690/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/106
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/106/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/106/comments
https://api.github.com/repos/ollama/ollama/issues/106/events
https://github.com/ollama/ollama/pull/106
1,810,840,013
PR_kwDOJ0Z1Ps5V1VZH
106
README typo fix
{ "login": "isaac-mcfadyen", "id": 6243993, "node_id": "MDQ6VXNlcjYyNDM5OTM=", "avatar_url": "https://avatars.githubusercontent.com/u/6243993?v=4", "gravatar_id": "", "url": "https://api.github.com/users/isaac-mcfadyen", "html_url": "https://github.com/isaac-mcfadyen", "followers_url": "https://api.github.com/users/isaac-mcfadyen/followers", "following_url": "https://api.github.com/users/isaac-mcfadyen/following{/other_user}", "gists_url": "https://api.github.com/users/isaac-mcfadyen/gists{/gist_id}", "starred_url": "https://api.github.com/users/isaac-mcfadyen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/isaac-mcfadyen/subscriptions", "organizations_url": "https://api.github.com/users/isaac-mcfadyen/orgs", "repos_url": "https://api.github.com/users/isaac-mcfadyen/repos", "events_url": "https://api.github.com/users/isaac-mcfadyen/events{/privacy}", "received_events_url": "https://api.github.com/users/isaac-mcfadyen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-07-18T23:00:59
2023-07-18T23:24:58
2023-07-18T23:24:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/106", "html_url": "https://github.com/ollama/ollama/pull/106", "diff_url": "https://github.com/ollama/ollama/pull/106.diff", "patch_url": "https://github.com/ollama/ollama/pull/106.patch", "merged_at": "2023-07-18T23:24:58" }
- Simple typo fix in README.md - `13` changed to `13B` (the `B` was missing) on Nous-Hermes model in models table - Also edited `hous-hermes` to `nous-hermes` (should be n instead of h)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/106/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/106/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3936
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3936/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3936/comments
https://api.github.com/repos/ollama/ollama/issues/3936/events
https://github.com/ollama/ollama/pull/3936
2,265,316,091
PR_kwDOJ0Z1Ps5t0VM1
3,936
pulled model not found fix - added to FAQ
{ "login": "lambdaofgod", "id": 3647577, "node_id": "MDQ6VXNlcjM2NDc1Nzc=", "avatar_url": "https://avatars.githubusercontent.com/u/3647577?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lambdaofgod", "html_url": "https://github.com/lambdaofgod", "followers_url": "https://api.github.com/users/lambdaofgod/followers", "following_url": "https://api.github.com/users/lambdaofgod/following{/other_user}", "gists_url": "https://api.github.com/users/lambdaofgod/gists{/gist_id}", "starred_url": "https://api.github.com/users/lambdaofgod/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lambdaofgod/subscriptions", "organizations_url": "https://api.github.com/users/lambdaofgod/orgs", "repos_url": "https://api.github.com/users/lambdaofgod/repos", "events_url": "https://api.github.com/users/lambdaofgod/events{/privacy}", "received_events_url": "https://api.github.com/users/lambdaofgod/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2024-04-26T09:04:22
2024-07-03T17:00:00
2024-07-03T16:59:59
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3936", "html_url": "https://github.com/ollama/ollama/pull/3936", "diff_url": "https://github.com/ollama/ollama/pull/3936.diff", "patch_url": "https://github.com/ollama/ollama/pull/3936.patch", "merged_at": null }
I've added a fix by [sridvijay](https://github.com/sridvijay) from #3876 to the documentation.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3936/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3936/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7864
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7864/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7864/comments
https://api.github.com/repos/ollama/ollama/issues/7864/events
https://github.com/ollama/ollama/issues/7864
2,698,937,631
I_kwDOJ0Z1Ps6g3oUf
7,864
Please sync with llama.cpp for the update of bert_base like models.
{ "login": "BeNhNp", "id": 33339730, "node_id": "MDQ6VXNlcjMzMzM5NzMw", "avatar_url": "https://avatars.githubusercontent.com/u/33339730?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BeNhNp", "html_url": "https://github.com/BeNhNp", "followers_url": "https://api.github.com/users/BeNhNp/followers", "following_url": "https://api.github.com/users/BeNhNp/following{/other_user}", "gists_url": "https://api.github.com/users/BeNhNp/gists{/gist_id}", "starred_url": "https://api.github.com/users/BeNhNp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BeNhNp/subscriptions", "organizations_url": "https://api.github.com/users/BeNhNp/orgs", "repos_url": "https://api.github.com/users/BeNhNp/repos", "events_url": "https://api.github.com/users/BeNhNp/events{/privacy}", "received_events_url": "https://api.github.com/users/BeNhNp/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
0
2024-11-27T15:34:28
2024-11-27T15:34:28
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? currently [ollama_llama_server](https://github.com/ollama/ollama/blob/main/llm/server.go#L894) can't return properly with google bert_base models, while llama.cpp have already supported, I've varified with `transformers`. ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7864/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7864/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3601
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3601/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3601/comments
https://api.github.com/repos/ollama/ollama/issues/3601/events
https://github.com/ollama/ollama/issues/3601
2,238,409,901
I_kwDOJ0Z1Ps6Fa2yt
3,601
Docker 0.1.31 the 2nd Ollama cannot use its designated GPU
{ "login": "ww2283", "id": 21956104, "node_id": "MDQ6VXNlcjIxOTU2MTA0", "avatar_url": "https://avatars.githubusercontent.com/u/21956104?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ww2283", "html_url": "https://github.com/ww2283", "followers_url": "https://api.github.com/users/ww2283/followers", "following_url": "https://api.github.com/users/ww2283/following{/other_user}", "gists_url": "https://api.github.com/users/ww2283/gists{/gist_id}", "starred_url": "https://api.github.com/users/ww2283/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ww2283/subscriptions", "organizations_url": "https://api.github.com/users/ww2283/orgs", "repos_url": "https://api.github.com/users/ww2283/repos", "events_url": "https://api.github.com/users/ww2283/events{/privacy}", "received_events_url": "https://api.github.com/users/ww2283/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-04-11T19:26:22
2024-04-11T19:36:45
2024-04-11T19:36:45
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm on Ubuntu 20.04 with two Ada6000 cards. I use docker compose to host two instance of Ollama, each has its own model because I want to use them for AutoGen. For each Ollama container I want to assign an individual GPU. In real use, I noticed that ollama1 is working perfectly fine, while ollama2 is working but only use CPU for inference. So I would appreciate the help to get the ollama2 to pick up its designated GPU unit. Here is my compose yaml file: ``` services: ollama1: image: ollama/ollama:latest container_name: ollama_model_1 ports: - '11435:11434' environment: - CUDA_VISIBLE_DEVICES=GPU-************************* deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] volumes: - /ADATAtmp/docker/ollama_model_1:/root/.ollama networks: - autogen_network restart: always open-webui1: image: ghcr.io/open-webui/open-webui:main container_name: open-webui_1 ports: - "11436:8080" volumes: - /ADATAtmp/docker/open-webui-1:/app/backend/data environment: - OLLAMA_BASE_URL=http://ollama_model_1:11434 extra_hosts: - "host.docker.internal:host-gateway" networks: - autogen_network restart: always ollama2: image: ollama/ollama:latest container_name: ollama_model_2 ports: - '11437:11434' environment: - CUDA_VISIBLE_DEVICES=GPU-************************* deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] volumes: - /ADATAtmp/docker/ollama_model_2:/root/.ollama networks: - autogen_network restart: always open-webui2: image: ghcr.io/open-webui/open-webui:main container_name: open-webui_2 ports: - "11438:8080" volumes: - /ADATAtmp/docker/open-webui-2:/app/backend/data environment: - OLLAMA_BASE_URL=http://ollama_model_2:11434 extra_hosts: - "host.docker.internal:host-gateway" networks: - autogen_network restart: always networks: autogen_network: driver: bridge ``` Also I believe this part of docker logs ollama2 is relevant for reviewing: ``` $ docker logs ollama_model_2 time=2024-04-11T19:05:33.375Z level=INFO source=images.go:804 msg="total blobs: 10" time=2024-04-11T19:05:33.375Z level=INFO source=images.go:811 msg="total unused blobs removed: 0" time=2024-04-11T19:05:33.375Z level=INFO source=routes.go:1118 msg="Listening on [::]:11434 (version 0.1.31)" time=2024-04-11T19:05:33.376Z level=INFO source=payload_common.go:113 msg="Extracting dynamic libraries to /tmp/ollama859516905/runners ..." time=2024-04-11T19:05:35.857Z level=INFO source=payload_common.go:140 msg="Dynamic LLM libraries [rocm_v60000 cpu cpu_avx2 cpu_avx cuda_v11]" time=2024-04-11T19:05:35.857Z level=INFO source=gpu.go:115 msg="Detecting GPU type" time=2024-04-11T19:05:35.857Z level=INFO source=gpu.go:265 msg="Searching for GPU management library libcudart.so*" time=2024-04-11T19:05:35.857Z level=INFO source=gpu.go:311 msg="Discovered GPU libraries: [/tmp/ollama859516905/runners/cuda_v11/libcudart.so.11.0]" time=2024-04-11T19:05:35.879Z level=INFO source=gpu.go:340 msg="Unable to load cudart CUDA management library /tmp/ollama859516905/runners/cuda_v11/libcudart.so.11.0: cudart init failure: 100" time=2024-04-11T19:05:35.879Z level=INFO source=gpu.go:265 msg="Searching for GPU management library libnvidia-ml.so" time=2024-04-11T19:05:35.879Z level=INFO source=gpu.go:311 msg="Discovered GPU libraries: [/usr/lib/x86_64-linux-gnu/libnvidia-ml.so.535.154.05]" time=2024-04-11T19:05:35.885Z level=INFO source=gpu.go:131 msg="Nvidia GPU detected via nvidia-ml" time=2024-04-11T19:05:35.885Z level=INFO source=cpu_common.go:11 msg="CPU has AVX2" time=2024-04-11T19:05:35.890Z level=INFO source=gpu.go:169 msg="[nvidia-ml] NVML CUDA Compute Capability detected: 8.9" [GIN] 2024/04/11 - 19:08:54 | 200 | 630.936µs | 172.20.0.5 | GET "/api/tags" [GIN] 2024/04/11 - 19:08:54 | 200 | 568.095µs | 172.20.0.5 | GET "/api/tags" [GIN] 2024/04/11 - 19:08:54 | 200 | 572.393µs | 172.20.0.5 | GET "/api/tags" [GIN] 2024/04/11 - 19:08:54 | 200 | 36.179µs | 172.20.0.5 | GET "/api/version" [GIN] 2024/04/11 - 19:08:56 | 200 | 38.734µs | 172.20.0.5 | GET "/api/version" time=2024-04-11T19:09:04.802Z level=INFO source=cpu_common.go:11 msg="CPU has AVX2" time=2024-04-11T19:09:04.802Z level=INFO source=gpu.go:169 msg="[nvidia-ml] NVML CUDA Compute Capability detected: 8.9" time=2024-04-11T19:09:04.802Z level=INFO source=cpu_common.go:11 msg="CPU has AVX2" time=2024-04-11T19:09:04.802Z level=INFO source=gpu.go:169 msg="[nvidia-ml] NVML CUDA Compute Capability detected: 8.9" time=2024-04-11T19:09:04.802Z level=INFO source=cpu_common.go:11 msg="CPU has AVX2" time=2024-04-11T19:09:04.807Z level=INFO source=dyn_ext_server.go:87 msg="Loading Dynamic llm server: /tmp/ollama859516905/runners/cuda_v11/libext_server.so" time=2024-04-11T19:09:04.807Z level=INFO source=dyn_ext_server.go:147 msg="Initializing llama server" time=2024-04-11T19:09:04.827Z level=WARN source=llm.go:170 msg="Failed to load dynamic library /tmp/ollama859516905/runners/cuda_v11/libext_server.so Unable to init GPU: no CUDA-capable device is detected" time=2024-04-11T19:09:04.828Z level=INFO source=dyn_ext_server.go:87 msg="Loading Dynamic llm server: /tmp/ollama859516905/runners/cpu_avx2/libext_server.so" time=2024-04-11T19:09:04.828Z level=INFO source=dyn_ext_server.go:147 msg="Initializing llama server" llama_model_loader: loaded meta data with 26 key-value pairs and 995 tensors from /root/.ollama/models/blobs/sha256-d68d6a65178011b746d215273d6a1f607f78be24a53532cf99618a32c2f382a2 (version GGUF V3 (latest)) # ...rest of the log ``` ### What did you expect to see? _No response_ ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS Linux ### Architecture x86 ### Platform Docker ### Ollama version 0.1.31 ### GPU Nvidia ### GPU info _No response_ ### CPU AMD ### Other software _No response_
{ "login": "ww2283", "id": 21956104, "node_id": "MDQ6VXNlcjIxOTU2MTA0", "avatar_url": "https://avatars.githubusercontent.com/u/21956104?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ww2283", "html_url": "https://github.com/ww2283", "followers_url": "https://api.github.com/users/ww2283/followers", "following_url": "https://api.github.com/users/ww2283/following{/other_user}", "gists_url": "https://api.github.com/users/ww2283/gists{/gist_id}", "starred_url": "https://api.github.com/users/ww2283/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ww2283/subscriptions", "organizations_url": "https://api.github.com/users/ww2283/orgs", "repos_url": "https://api.github.com/users/ww2283/repos", "events_url": "https://api.github.com/users/ww2283/events{/privacy}", "received_events_url": "https://api.github.com/users/ww2283/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3601/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3601/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6696
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6696/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6696/comments
https://api.github.com/repos/ollama/ollama/issues/6696/events
https://github.com/ollama/ollama/pull/6696
2,512,217,526
PR_kwDOJ0Z1Ps56wsWl
6,696
Update README.md
{ "login": "rapidarchitect", "id": 126218667, "node_id": "U_kgDOB4Xxqw", "avatar_url": "https://avatars.githubusercontent.com/u/126218667?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rapidarchitect", "html_url": "https://github.com/rapidarchitect", "followers_url": "https://api.github.com/users/rapidarchitect/followers", "following_url": "https://api.github.com/users/rapidarchitect/following{/other_user}", "gists_url": "https://api.github.com/users/rapidarchitect/gists{/gist_id}", "starred_url": "https://api.github.com/users/rapidarchitect/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rapidarchitect/subscriptions", "organizations_url": "https://api.github.com/users/rapidarchitect/orgs", "repos_url": "https://api.github.com/users/rapidarchitect/repos", "events_url": "https://api.github.com/users/rapidarchitect/events{/privacy}", "received_events_url": "https://api.github.com/users/rapidarchitect/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-09-08T06:08:35
2024-09-08T07:36:00
2024-09-08T07:36:00
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6696", "html_url": "https://github.com/ollama/ollama/pull/6696", "diff_url": "https://github.com/ollama/ollama/pull/6696.diff", "patch_url": "https://github.com/ollama/ollama/pull/6696.patch", "merged_at": "2024-09-08T07:36:00" }
added crewai with mesop anew since there was a conflict previously
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6696/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6696/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6029
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6029/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6029/comments
https://api.github.com/repos/ollama/ollama/issues/6029/events
https://github.com/ollama/ollama/issues/6029
2,434,070,538
I_kwDOJ0Z1Ps6RFPgK
6,029
Prompt evaluation progress indicator
{ "login": "drazdra", "id": 133811709, "node_id": "U_kgDOB_nN_Q", "avatar_url": "https://avatars.githubusercontent.com/u/133811709?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drazdra", "html_url": "https://github.com/drazdra", "followers_url": "https://api.github.com/users/drazdra/followers", "following_url": "https://api.github.com/users/drazdra/following{/other_user}", "gists_url": "https://api.github.com/users/drazdra/gists{/gist_id}", "starred_url": "https://api.github.com/users/drazdra/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drazdra/subscriptions", "organizations_url": "https://api.github.com/users/drazdra/orgs", "repos_url": "https://api.github.com/users/drazdra/repos", "events_url": "https://api.github.com/users/drazdra/events{/privacy}", "received_events_url": "https://api.github.com/users/drazdra/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
1
2024-07-28T19:05:02
2024-08-14T05:23:01
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Prompt evaluation can take huge time, especially with long context. it can literary be hours. Right now Ollama just hangs in that phase. Is it possible to have some messages updating on the progress? It's tough when after an hour of 100% load you still don't know if it's at 15% or at 99%..
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6029/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6029/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/823
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/823/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/823/comments
https://api.github.com/repos/ollama/ollama/issues/823/events
https://github.com/ollama/ollama/issues/823
1,948,000,329
I_kwDOJ0Z1Ps50HCBJ
823
failed to verify certificate: x509: certificate signed by unknown authority
{ "login": "sureshpatel66", "id": 87146534, "node_id": "MDQ6VXNlcjg3MTQ2NTM0", "avatar_url": "https://avatars.githubusercontent.com/u/87146534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sureshpatel66", "html_url": "https://github.com/sureshpatel66", "followers_url": "https://api.github.com/users/sureshpatel66/followers", "following_url": "https://api.github.com/users/sureshpatel66/following{/other_user}", "gists_url": "https://api.github.com/users/sureshpatel66/gists{/gist_id}", "starred_url": "https://api.github.com/users/sureshpatel66/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sureshpatel66/subscriptions", "organizations_url": "https://api.github.com/users/sureshpatel66/orgs", "repos_url": "https://api.github.com/users/sureshpatel66/repos", "events_url": "https://api.github.com/users/sureshpatel66/events{/privacy}", "received_events_url": "https://api.github.com/users/sureshpatel66/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
18
2023-10-17T18:21:33
2024-11-21T10:18:36
2023-10-25T19:10:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
~$ docker exec -it ollama ollama run llama2 pulling manifest Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": tls: failed to verify certificate: x509: certificate signed by unknown authority please guide to solve this issue
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/823/reactions", "total_count": 7, "+1": 7, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/823/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/7281
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7281/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7281/comments
https://api.github.com/repos/ollama/ollama/issues/7281/events
https://github.com/ollama/ollama/pull/7281
2,600,936,244
PR_kwDOJ0Z1Ps5_PgOY
7,281
added formatting to ollama ls
{ "login": "robbiemu", "id": 248927, "node_id": "MDQ6VXNlcjI0ODkyNw==", "avatar_url": "https://avatars.githubusercontent.com/u/248927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/robbiemu", "html_url": "https://github.com/robbiemu", "followers_url": "https://api.github.com/users/robbiemu/followers", "following_url": "https://api.github.com/users/robbiemu/following{/other_user}", "gists_url": "https://api.github.com/users/robbiemu/gists{/gist_id}", "starred_url": "https://api.github.com/users/robbiemu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/robbiemu/subscriptions", "organizations_url": "https://api.github.com/users/robbiemu/orgs", "repos_url": "https://api.github.com/users/robbiemu/repos", "events_url": "https://api.github.com/users/robbiemu/events{/privacy}", "received_events_url": "https://api.github.com/users/robbiemu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
1
2024-10-20T22:44:50
2024-11-06T22:54:30
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7281", "html_url": "https://github.com/ollama/ollama/pull/7281", "diff_url": "https://github.com/ollama/ollama/pull/7281.diff", "patch_url": "https://github.com/ollama/ollama/pull/7281.patch", "merged_at": null }
This groups output by model on disk, and does not pollute the filesize in the listing when the manifests point to the same model. ``` ./ollama ls NAME ID SIZE MODIFIED qwen2.5:32b-instruct-q4_K_M 9f13ba1299af 19 GB 9 hours ago bge-m3:latest 790764642607 1.2 GB 3 weeks ago llama3.2:3b-instruct-q8_0 e410b836fe61 3.4 GB 3 weeks ago reader-lm:1.5b-q8_0 39ec9eda0492 1.6 GB 4 weeks ago └── reader-lm:1.5b-q8_0_defaults 05fe7cfa3bea -- 3 days ago gemma2:27b-instruct-q6_K 73a0cbed4b12 22 GB 3 months ago ``` Caveats - it seems the "basemodel" is not distinguished in the data except by modifiedBy -- which perhaps could be updated by the user without deleting or updating manifests they created (these descendent manifests are where you use "/set PARAMETER ..." and "/save ..."). - I had to explicitly turn off AutoWrapText because it was wrapping in the column at the space around the prefixes (like "└── "), and setting min column width just for this model does not effect this wrapping behavior.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7281/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7281/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1685
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1685/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1685/comments
https://api.github.com/repos/ollama/ollama/issues/1685/events
https://github.com/ollama/ollama/issues/1685
2,054,605,354
I_kwDOJ0Z1Ps56dsoq
1,685
Ollama mobile app appeared on iOS app store
{ "login": "netsap", "id": 41242395, "node_id": "MDQ6VXNlcjQxMjQyMzk1", "avatar_url": "https://avatars.githubusercontent.com/u/41242395?v=4", "gravatar_id": "", "url": "https://api.github.com/users/netsap", "html_url": "https://github.com/netsap", "followers_url": "https://api.github.com/users/netsap/followers", "following_url": "https://api.github.com/users/netsap/following{/other_user}", "gists_url": "https://api.github.com/users/netsap/gists{/gist_id}", "starred_url": "https://api.github.com/users/netsap/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/netsap/subscriptions", "organizations_url": "https://api.github.com/users/netsap/orgs", "repos_url": "https://api.github.com/users/netsap/repos", "events_url": "https://api.github.com/users/netsap/events{/privacy}", "received_events_url": "https://api.github.com/users/netsap/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2023-12-23T01:05:23
2024-05-10T00:27:44
2024-05-10T00:27:44
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Accessible here: https://apps.apple.com/us/app/ollama/id6471840114 I'm assuming this is not legitimate?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1685/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 2, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1685/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2043
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2043/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2043/comments
https://api.github.com/repos/ollama/ollama/issues/2043/events
https://github.com/ollama/ollama/pull/2043
2,087,600,434
PR_kwDOJ0Z1Ps5kY2C0
2,043
Dockerfile: use variables for package version
{ "login": "stevenbecht", "id": 9442836, "node_id": "MDQ6VXNlcjk0NDI4MzY=", "avatar_url": "https://avatars.githubusercontent.com/u/9442836?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevenbecht", "html_url": "https://github.com/stevenbecht", "followers_url": "https://api.github.com/users/stevenbecht/followers", "following_url": "https://api.github.com/users/stevenbecht/following{/other_user}", "gists_url": "https://api.github.com/users/stevenbecht/gists{/gist_id}", "starred_url": "https://api.github.com/users/stevenbecht/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevenbecht/subscriptions", "organizations_url": "https://api.github.com/users/stevenbecht/orgs", "repos_url": "https://api.github.com/users/stevenbecht/repos", "events_url": "https://api.github.com/users/stevenbecht/events{/privacy}", "received_events_url": "https://api.github.com/users/stevenbecht/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-18T06:23:39
2024-02-21T00:26:49
2024-02-21T00:26:49
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2043", "html_url": "https://github.com/ollama/ollama/pull/2043", "diff_url": "https://github.com/ollama/ollama/pull/2043.diff", "patch_url": "https://github.com/ollama/ollama/pull/2043.patch", "merged_at": null }
Update Dockerfile to use variables instead of hardcoded values
{ "login": "stevenbecht", "id": 9442836, "node_id": "MDQ6VXNlcjk0NDI4MzY=", "avatar_url": "https://avatars.githubusercontent.com/u/9442836?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stevenbecht", "html_url": "https://github.com/stevenbecht", "followers_url": "https://api.github.com/users/stevenbecht/followers", "following_url": "https://api.github.com/users/stevenbecht/following{/other_user}", "gists_url": "https://api.github.com/users/stevenbecht/gists{/gist_id}", "starred_url": "https://api.github.com/users/stevenbecht/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevenbecht/subscriptions", "organizations_url": "https://api.github.com/users/stevenbecht/orgs", "repos_url": "https://api.github.com/users/stevenbecht/repos", "events_url": "https://api.github.com/users/stevenbecht/events{/privacy}", "received_events_url": "https://api.github.com/users/stevenbecht/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2043/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2043/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5285
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5285/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5285/comments
https://api.github.com/repos/ollama/ollama/issues/5285/events
https://github.com/ollama/ollama/pull/5285
2,373,797,637
PR_kwDOJ0Z1Ps5zjnWI
5,285
OpenAI: /v1/embeddings compatibility
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-06-25T22:47:36
2024-07-16T20:36:11
2024-07-16T20:36:09
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5285", "html_url": "https://github.com/ollama/ollama/pull/5285", "diff_url": "https://github.com/ollama/ollama/pull/5285.diff", "patch_url": "https://github.com/ollama/ollama/pull/5285.patch", "merged_at": "2024-07-16T20:36:08" }
In anticipation of #5127 Resolves #2416 ``` curl http://localhost:11434/v1/embeddings \ -H "Content-Type: application/json" \ -d '{ "input": ["hello", "hi"], "model": "all-minilm" }' ``` ``` { "object": "list", "data": [ { "object": "embedding", "embedding": [ ... ], "index": 0 }, { "object": "embedding", "embedding": [ ... ], "index": 1 } ], "model": "all-minilm" } ```
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5285/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 2, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/5285/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7969
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7969/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7969/comments
https://api.github.com/repos/ollama/ollama/issues/7969/events
https://github.com/ollama/ollama/issues/7969
2,723,362,146
I_kwDOJ0Z1Ps6iUzVi
7,969
Administrative / silent install is borked
{ "login": "Lukas-UAUX", "id": 43141094, "node_id": "MDQ6VXNlcjQzMTQxMDk0", "avatar_url": "https://avatars.githubusercontent.com/u/43141094?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Lukas-UAUX", "html_url": "https://github.com/Lukas-UAUX", "followers_url": "https://api.github.com/users/Lukas-UAUX/followers", "following_url": "https://api.github.com/users/Lukas-UAUX/following{/other_user}", "gists_url": "https://api.github.com/users/Lukas-UAUX/gists{/gist_id}", "starred_url": "https://api.github.com/users/Lukas-UAUX/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Lukas-UAUX/subscriptions", "organizations_url": "https://api.github.com/users/Lukas-UAUX/orgs", "repos_url": "https://api.github.com/users/Lukas-UAUX/repos", "events_url": "https://api.github.com/users/Lukas-UAUX/events{/privacy}", "received_events_url": "https://api.github.com/users/Lukas-UAUX/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" }, { "id": 6678628138, "node_id": "LA_kwDOJ0Z1Ps8AAAABjhPHKg", "url": "https://api.github.com/repos/ollama/ollama/labels/install", "name": "install", "color": "E0B88D", "default": false, "description": "" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-12-06T15:50:06
2024-12-10T17:30:35
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? For deployment scenarios like for classrooms or to regular managed devices the setup needs to perform an administrative / unattended install. For InnoSetup built installers this can normally be done via the command-line switches /SILENT (or /VERYSILENT) or via a response file via /LOADINF. (See: https://jrsoftware.org/ishelp/index.php?topic=setupcmdline ) If the setup can be performed by both, an unprivileged user and an administrator the switch "/ALLUSERS" helps to select the latter. If the install paths are (correctly) derived from the common environment variables this should automagically move the files to "C:\Program Files" or "C:\Program Files (x86)" instead of the users AppData folder. Normally this would also lead the installer engine to register the application under the systems registry (HKLM:) and create a system wide start menu. With Ollamas setup this does not work even though setting the install-path like mentioned in the documentation to 'C:\Program Files\Ollama' makes it look like it should. Instead it creates a mix of both methods where the files are installed to the Program Files folder, but the registry and start menu entries are limited to the user (context) installing the software. And as most management system use a local system service as context this creates a rather weird state. And even if I could limit the permissions down - for a classroom for example I can never know which user will access the device... ### OS Windows ### GPU _No response_ ### CPU _No response_ ### Ollama version 0.4.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7969/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7969/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2811
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2811/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2811/comments
https://api.github.com/repos/ollama/ollama/issues/2811/events
https://github.com/ollama/ollama/issues/2811
2,159,235,157
I_kwDOJ0Z1Ps6As1BV
2,811
Mistral Instruct models prompt does not use <s> or </s>
{ "login": "louisabraham", "id": 13174805, "node_id": "MDQ6VXNlcjEzMTc0ODA1", "avatar_url": "https://avatars.githubusercontent.com/u/13174805?v=4", "gravatar_id": "", "url": "https://api.github.com/users/louisabraham", "html_url": "https://github.com/louisabraham", "followers_url": "https://api.github.com/users/louisabraham/followers", "following_url": "https://api.github.com/users/louisabraham/following{/other_user}", "gists_url": "https://api.github.com/users/louisabraham/gists{/gist_id}", "starred_url": "https://api.github.com/users/louisabraham/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/louisabraham/subscriptions", "organizations_url": "https://api.github.com/users/louisabraham/orgs", "repos_url": "https://api.github.com/users/louisabraham/repos", "events_url": "https://api.github.com/users/louisabraham/events{/privacy}", "received_events_url": "https://api.github.com/users/louisabraham/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
5
2024-02-28T15:29:34
2024-07-18T22:45:02
2024-07-18T22:45:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
from https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1#instruction-format > \<s> [INST] Instruction [/INST] Model answer\</s> [INST] Follow-up instruction [/INST] I didn't see `<s>` as part of the prompt when using `OLLAMA_DEBUG=1 ollama start`. Is that expected?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2811/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2811/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6611
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6611/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6611/comments
https://api.github.com/repos/ollama/ollama/issues/6611/events
https://github.com/ollama/ollama/pull/6611
2,503,463,117
PR_kwDOJ0Z1Ps56S5Sr
6,611
Make stall duration timeout configurable
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-09-03T18:04:25
2024-09-05T21:00:08
2024-09-05T21:00:08
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6611", "html_url": "https://github.com/ollama/ollama/pull/6611", "diff_url": "https://github.com/ollama/ollama/pull/6611.diff", "patch_url": "https://github.com/ollama/ollama/pull/6611.patch", "merged_at": "2024-09-05T21:00:08" }
With the new very large parameter models, some users are willing to wait for a very long time for models to load. Fixes #6031
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6611/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6611/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/329
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/329/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/329/comments
https://api.github.com/repos/ollama/ollama/issues/329/events
https://github.com/ollama/ollama/pull/329
1,846,179,798
PR_kwDOJ0Z1Ps5XsTL-
329
Add tutorials for using Langchain with ollama
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-08-11T04:30:50
2024-01-05T18:40:54
2023-08-11T22:19:39
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/329", "html_url": "https://github.com/ollama/ollama/pull/329", "diff_url": "https://github.com/ollama/ollama/pull/329.diff", "patch_url": "https://github.com/ollama/ollama/pull/329.patch", "merged_at": "2023-08-11T22:19:39" }
null
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/329/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/329/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7241
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7241/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7241/comments
https://api.github.com/repos/ollama/ollama/issues/7241/events
https://github.com/ollama/ollama/issues/7241
2,595,035,489
I_kwDOJ0Z1Ps6arRlh
7,241
add module/
{ "login": "malv-c", "id": 19170213, "node_id": "MDQ6VXNlcjE5MTcwMjEz", "avatar_url": "https://avatars.githubusercontent.com/u/19170213?v=4", "gravatar_id": "", "url": "https://api.github.com/users/malv-c", "html_url": "https://github.com/malv-c", "followers_url": "https://api.github.com/users/malv-c/followers", "following_url": "https://api.github.com/users/malv-c/following{/other_user}", "gists_url": "https://api.github.com/users/malv-c/gists{/gist_id}", "starred_url": "https://api.github.com/users/malv-c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/malv-c/subscriptions", "organizations_url": "https://api.github.com/users/malv-c/orgs", "repos_url": "https://api.github.com/users/malv-c/repos", "events_url": "https://api.github.com/users/malv-c/events{/privacy}", "received_events_url": "https://api.github.com/users/malv-c/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-10-17T15:12:19
2024-10-20T06:06:27
2024-10-20T06:06:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "malv-c", "id": 19170213, "node_id": "MDQ6VXNlcjE5MTcwMjEz", "avatar_url": "https://avatars.githubusercontent.com/u/19170213?v=4", "gravatar_id": "", "url": "https://api.github.com/users/malv-c", "html_url": "https://github.com/malv-c", "followers_url": "https://api.github.com/users/malv-c/followers", "following_url": "https://api.github.com/users/malv-c/following{/other_user}", "gists_url": "https://api.github.com/users/malv-c/gists{/gist_id}", "starred_url": "https://api.github.com/users/malv-c/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/malv-c/subscriptions", "organizations_url": "https://api.github.com/users/malv-c/orgs", "repos_url": "https://api.github.com/users/malv-c/repos", "events_url": "https://api.github.com/users/malv-c/events{/privacy}", "received_events_url": "https://api.github.com/users/malv-c/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7241/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7241/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2602
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2602/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2602/comments
https://api.github.com/repos/ollama/ollama/issues/2602/events
https://github.com/ollama/ollama/issues/2602
2,143,418,344
I_kwDOJ0Z1Ps5_wffo
2,602
Windows preview - please let us set the location where models are stored
{ "login": "axefrog", "id": 298883, "node_id": "MDQ6VXNlcjI5ODg4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/298883?v=4", "gravatar_id": "", "url": "https://api.github.com/users/axefrog", "html_url": "https://github.com/axefrog", "followers_url": "https://api.github.com/users/axefrog/followers", "following_url": "https://api.github.com/users/axefrog/following{/other_user}", "gists_url": "https://api.github.com/users/axefrog/gists{/gist_id}", "starred_url": "https://api.github.com/users/axefrog/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/axefrog/subscriptions", "organizations_url": "https://api.github.com/users/axefrog/orgs", "repos_url": "https://api.github.com/users/axefrog/repos", "events_url": "https://api.github.com/users/axefrog/events{/privacy}", "received_events_url": "https://api.github.com/users/axefrog/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-02-20T02:36:27
2024-03-06T18:06:53
2024-02-20T03:38:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
My drive C is a bit ancient right now. It's an old 250GB SSD and at any given time seems to have about 5-10gb free, so I'm forever doing cleanups to stop it running out of space. In contrast, I have about 2-3TB of free space on my other drives. I would _much_ prefer it if ollama would store models on one of those drives. How much effort would it be to make that happen?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2602/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2602/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7647
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7647/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7647/comments
https://api.github.com/repos/ollama/ollama/issues/7647/events
https://github.com/ollama/ollama/issues/7647
2,654,957,622
I_kwDOJ0Z1Ps6eP3A2
7,647
Sending "hidden" messages for attachments
{ "login": "robotom", "id": 45123215, "node_id": "MDQ6VXNlcjQ1MTIzMjE1", "avatar_url": "https://avatars.githubusercontent.com/u/45123215?v=4", "gravatar_id": "", "url": "https://api.github.com/users/robotom", "html_url": "https://github.com/robotom", "followers_url": "https://api.github.com/users/robotom/followers", "following_url": "https://api.github.com/users/robotom/following{/other_user}", "gists_url": "https://api.github.com/users/robotom/gists{/gist_id}", "starred_url": "https://api.github.com/users/robotom/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/robotom/subscriptions", "organizations_url": "https://api.github.com/users/robotom/orgs", "repos_url": "https://api.github.com/users/robotom/repos", "events_url": "https://api.github.com/users/robotom/events{/privacy}", "received_events_url": "https://api.github.com/users/robotom/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-11-13T10:20:06
2024-11-13T19:53:25
2024-11-13T19:53:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm building a frontend and trying to send attachment contents and an accompanying prompt (e.g. "analyze this and wait for user prompt before responding") as a "hidden" message so that it's not rendered to the user on the frontend. I'm trying to avoid seeing 10 pages of a doc rendered in the chat window. Perhaps this is a frontend issue but I'm finding this more difficult to code than I think it should be... Any ideas? Thanks!
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7647/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7647/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8349
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8349/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8349/comments
https://api.github.com/repos/ollama/ollama/issues/8349/events
https://github.com/ollama/ollama/issues/8349
2,775,993,379
I_kwDOJ0Z1Ps6ldkwj
8,349
can't use gpu after update
{ "login": "Heart-eartH", "id": 75364983, "node_id": "MDQ6VXNlcjc1MzY0OTgz", "avatar_url": "https://avatars.githubusercontent.com/u/75364983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Heart-eartH", "html_url": "https://github.com/Heart-eartH", "followers_url": "https://api.github.com/users/Heart-eartH/followers", "following_url": "https://api.github.com/users/Heart-eartH/following{/other_user}", "gists_url": "https://api.github.com/users/Heart-eartH/gists{/gist_id}", "starred_url": "https://api.github.com/users/Heart-eartH/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Heart-eartH/subscriptions", "organizations_url": "https://api.github.com/users/Heart-eartH/orgs", "repos_url": "https://api.github.com/users/Heart-eartH/repos", "events_url": "https://api.github.com/users/Heart-eartH/events{/privacy}", "received_events_url": "https://api.github.com/users/Heart-eartH/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" }, { "id": 6677745918, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g", "url": "https://api.github.com/repos/ollama/ollama/labels/gpu", "name": "gpu", "color": "76C49E", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
7
2025-01-08T17:57:28
2025-01-28T20:46:29
2025-01-28T20:46:29
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Previously, it run on the GPU by default, but after the update, it run on the CPU. What can I do to change it ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.4
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8349/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8349/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1552
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1552/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1552/comments
https://api.github.com/repos/ollama/ollama/issues/1552/events
https://github.com/ollama/ollama/pull/1552
2,044,274,427
PR_kwDOJ0Z1Ps5iI-Kk
1,552
add lint and test on pull_request
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-15T19:34:07
2024-01-11T17:37:47
2024-01-11T17:37:46
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1552", "html_url": "https://github.com/ollama/ollama/pull/1552", "diff_url": "https://github.com/ollama/ollama/pull/1552.diff", "patch_url": "https://github.com/ollama/ollama/pull/1552.patch", "merged_at": "2024-01-11T17:37:46" }
fixes a bug with generate where `get_flags` errors on ubuntu (and likely windows) when building cuda on a cuda-less system fixes a bug in windows where `/api/list` does not return models correctly both lint and test requires go generate results so do it once then propagate the artifacts to the rest of the pipeline this enables linting with golangci-lint but doesn't go overboard with linters. it enables the default linters and only a few extra linters to catch the most egregious bugs linting can run locally with [golangci-lint](https://golangci-lint.run/usage/install/) ``` go install github.com/golangci/golangci-lint/cmd/golangci-lint@v1.55.2 golangci-lint run -v ``` ``` docker run --rm -v $(pwd):/app -w /app golangci/golangci-lint:v1.55.2 golangci-lint run -v ``` resolves #1539
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1552/reactions", "total_count": 2, "+1": 1, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1552/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6272
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6272/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6272/comments
https://api.github.com/repos/ollama/ollama/issues/6272/events
https://github.com/ollama/ollama/issues/6272
2,457,035,391
I_kwDOJ0Z1Ps6Sc2J_
6,272
Ollama Creat 手动部署 报错 Error: invalid file magic
{ "login": "JaminYan", "id": 83393121, "node_id": "MDQ6VXNlcjgzMzkzMTIx", "avatar_url": "https://avatars.githubusercontent.com/u/83393121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JaminYan", "html_url": "https://github.com/JaminYan", "followers_url": "https://api.github.com/users/JaminYan/followers", "following_url": "https://api.github.com/users/JaminYan/following{/other_user}", "gists_url": "https://api.github.com/users/JaminYan/gists{/gist_id}", "starred_url": "https://api.github.com/users/JaminYan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JaminYan/subscriptions", "organizations_url": "https://api.github.com/users/JaminYan/orgs", "repos_url": "https://api.github.com/users/JaminYan/repos", "events_url": "https://api.github.com/users/JaminYan/events{/privacy}", "received_events_url": "https://api.github.com/users/JaminYan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
22
2024-08-09T03:47:19
2024-11-11T09:41:36
2024-08-25T20:15:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? 按飞书文档找的modelfile 文件,ollama creat 后输出报错 Error: invalid file magic ,因此无法部署在ollama ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.34
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6272/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6272/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5633
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5633/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5633/comments
https://api.github.com/repos/ollama/ollama/issues/5633/events
https://github.com/ollama/ollama/pull/5633
2,403,487,442
PR_kwDOJ0Z1Ps51HNIc
5,633
Update README.md >> Extension & Plugins :Terraform AWS Ollama & Open WebUI
{ "login": "xuyangbocn", "id": 11894567, "node_id": "MDQ6VXNlcjExODk0NTY3", "avatar_url": "https://avatars.githubusercontent.com/u/11894567?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xuyangbocn", "html_url": "https://github.com/xuyangbocn", "followers_url": "https://api.github.com/users/xuyangbocn/followers", "following_url": "https://api.github.com/users/xuyangbocn/following{/other_user}", "gists_url": "https://api.github.com/users/xuyangbocn/gists{/gist_id}", "starred_url": "https://api.github.com/users/xuyangbocn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/xuyangbocn/subscriptions", "organizations_url": "https://api.github.com/users/xuyangbocn/orgs", "repos_url": "https://api.github.com/users/xuyangbocn/repos", "events_url": "https://api.github.com/users/xuyangbocn/events{/privacy}", "received_events_url": "https://api.github.com/users/xuyangbocn/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-11T15:29:25
2024-11-21T10:28:58
2024-11-21T10:28:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5633", "html_url": "https://github.com/ollama/ollama/pull/5633", "diff_url": "https://github.com/ollama/ollama/pull/5633.diff", "patch_url": "https://github.com/ollama/ollama/pull/5633.patch", "merged_at": "2024-11-21T10:28:58" }
A Terraform module to deploy on AWS a ready-to-use Ollama service, together with its front end Open WebUI service.
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5633/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5633/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3733
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3733/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3733/comments
https://api.github.com/repos/ollama/ollama/issues/3733/events
https://github.com/ollama/ollama/issues/3733
2,250,701,058
I_kwDOJ0Z1Ps6GJvkC
3,733
I wrote an uninstallation script.
{ "login": "TikaFlow", "id": 33306473, "node_id": "MDQ6VXNlcjMzMzA2NDcz", "avatar_url": "https://avatars.githubusercontent.com/u/33306473?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TikaFlow", "html_url": "https://github.com/TikaFlow", "followers_url": "https://api.github.com/users/TikaFlow/followers", "following_url": "https://api.github.com/users/TikaFlow/following{/other_user}", "gists_url": "https://api.github.com/users/TikaFlow/gists{/gist_id}", "starred_url": "https://api.github.com/users/TikaFlow/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TikaFlow/subscriptions", "organizations_url": "https://api.github.com/users/TikaFlow/orgs", "repos_url": "https://api.github.com/users/TikaFlow/repos", "events_url": "https://api.github.com/users/TikaFlow/events{/privacy}", "received_events_url": "https://api.github.com/users/TikaFlow/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-04-18T13:26:28
2024-04-19T01:59:32
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I wrote an uninstallation script and [this](https://github.com/TikaFlow/tika_grocery/tree/main/delOllama) is my script. If you think it's not bad, you might consider referencing my script in your [documentation](https://github.com/ollama/ollama/blob/main/docs/linux.md#uninstall). Thanks!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3733/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3733/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4409
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4409/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4409/comments
https://api.github.com/repos/ollama/ollama/issues/4409/events
https://github.com/ollama/ollama/issues/4409
2,293,631,935
I_kwDOJ0Z1Ps6Itgu_
4,409
Windows security: ALERT: Trojan:Win32/Wacatac.H!ml detected
{ "login": "boessu", "id": 2807976, "node_id": "MDQ6VXNlcjI4MDc5NzY=", "avatar_url": "https://avatars.githubusercontent.com/u/2807976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/boessu", "html_url": "https://github.com/boessu", "followers_url": "https://api.github.com/users/boessu/followers", "following_url": "https://api.github.com/users/boessu/following{/other_user}", "gists_url": "https://api.github.com/users/boessu/gists{/gist_id}", "starred_url": "https://api.github.com/users/boessu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/boessu/subscriptions", "organizations_url": "https://api.github.com/users/boessu/orgs", "repos_url": "https://api.github.com/users/boessu/repos", "events_url": "https://api.github.com/users/boessu/events{/privacy}", "received_events_url": "https://api.github.com/users/boessu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-05-13T19:17:50
2024-05-13T20:03:24
2024-05-13T20:03:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? There is an alarm in Windows Defender which sets the following file under quarantine: file: %USERPROFILE%\AppData\Local\Programs\Ollama\ollama app.exe Windows Version: Windows 11 Pro, 23H2 You'll get the warning with update and also if you install it new. ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version 0.1.37
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4409/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4409/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5743
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5743/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5743/comments
https://api.github.com/repos/ollama/ollama/issues/5743/events
https://github.com/ollama/ollama/issues/5743
2,413,191,986
I_kwDOJ0Z1Ps6P1mMy
5,743
process cant't kill on windows 10
{ "login": "wszgrcy", "id": 9607121, "node_id": "MDQ6VXNlcjk2MDcxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/9607121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wszgrcy", "html_url": "https://github.com/wszgrcy", "followers_url": "https://api.github.com/users/wszgrcy/followers", "following_url": "https://api.github.com/users/wszgrcy/following{/other_user}", "gists_url": "https://api.github.com/users/wszgrcy/gists{/gist_id}", "starred_url": "https://api.github.com/users/wszgrcy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wszgrcy/subscriptions", "organizations_url": "https://api.github.com/users/wszgrcy/orgs", "repos_url": "https://api.github.com/users/wszgrcy/repos", "events_url": "https://api.github.com/users/wszgrcy/events{/privacy}", "received_events_url": "https://api.github.com/users/wszgrcy/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
5
2024-07-17T10:04:20
2024-07-19T03:29:24
2024-07-19T03:29:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I use child process in node to start ollama,and when I kill process,The process will be partially retained(`ollama.exe` is killed ollama_llama_server retained) ![image](https://github.com/user-attachments/assets/48d9f7d5-92dc-4357-ac1b-f293b994e788) Ubuntu will correctly kill all processes ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version 0.2.5
{ "login": "wszgrcy", "id": 9607121, "node_id": "MDQ6VXNlcjk2MDcxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/9607121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wszgrcy", "html_url": "https://github.com/wszgrcy", "followers_url": "https://api.github.com/users/wszgrcy/followers", "following_url": "https://api.github.com/users/wszgrcy/following{/other_user}", "gists_url": "https://api.github.com/users/wszgrcy/gists{/gist_id}", "starred_url": "https://api.github.com/users/wszgrcy/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wszgrcy/subscriptions", "organizations_url": "https://api.github.com/users/wszgrcy/orgs", "repos_url": "https://api.github.com/users/wszgrcy/repos", "events_url": "https://api.github.com/users/wszgrcy/events{/privacy}", "received_events_url": "https://api.github.com/users/wszgrcy/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5743/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5743/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7662
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7662/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7662/comments
https://api.github.com/repos/ollama/ollama/issues/7662/events
https://github.com/ollama/ollama/issues/7662
2,657,523,102
I_kwDOJ0Z1Ps6eZpWe
7,662
How can I remove specific PARAMETER from model?
{ "login": "wwjCMP", "id": 32979859, "node_id": "MDQ6VXNlcjMyOTc5ODU5", "avatar_url": "https://avatars.githubusercontent.com/u/32979859?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wwjCMP", "html_url": "https://github.com/wwjCMP", "followers_url": "https://api.github.com/users/wwjCMP/followers", "following_url": "https://api.github.com/users/wwjCMP/following{/other_user}", "gists_url": "https://api.github.com/users/wwjCMP/gists{/gist_id}", "starred_url": "https://api.github.com/users/wwjCMP/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wwjCMP/subscriptions", "organizations_url": "https://api.github.com/users/wwjCMP/orgs", "repos_url": "https://api.github.com/users/wwjCMP/repos", "events_url": "https://api.github.com/users/wwjCMP/events{/privacy}", "received_events_url": "https://api.github.com/users/wwjCMP/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-11-14T04:27:14
2024-11-14T14:17:11
2024-11-14T14:17:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How can I remove specific PARAMETER from model For example, delete PARAMETER num_gpu 30 from qwen2.5-coder:32b-instruct-q8_0 ollama show qwen2.5-coder:32b-instruct-q8_0 Model architecture qwen2 parameters 32.8B context length 32768 embedding length 5120 quantization Q8_0 Parameters num_ctx 16000 num_gpu 30 System You are Qwen, created by Alibaba Cloud. You are a helpful assistant. License Apache License Version 2.0, January 2004
{ "login": "wwjCMP", "id": 32979859, "node_id": "MDQ6VXNlcjMyOTc5ODU5", "avatar_url": "https://avatars.githubusercontent.com/u/32979859?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wwjCMP", "html_url": "https://github.com/wwjCMP", "followers_url": "https://api.github.com/users/wwjCMP/followers", "following_url": "https://api.github.com/users/wwjCMP/following{/other_user}", "gists_url": "https://api.github.com/users/wwjCMP/gists{/gist_id}", "starred_url": "https://api.github.com/users/wwjCMP/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wwjCMP/subscriptions", "organizations_url": "https://api.github.com/users/wwjCMP/orgs", "repos_url": "https://api.github.com/users/wwjCMP/repos", "events_url": "https://api.github.com/users/wwjCMP/events{/privacy}", "received_events_url": "https://api.github.com/users/wwjCMP/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7662/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7662/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3330
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3330/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3330/comments
https://api.github.com/repos/ollama/ollama/issues/3330/events
https://github.com/ollama/ollama/issues/3330
2,204,580,969
I_kwDOJ0Z1Ps6DZzxp
3,330
Failed to open the file /tmp/tmp.RbBP0lFvPD/ollama:
{ "login": "hansaskov", "id": 70580312, "node_id": "MDQ6VXNlcjcwNTgwMzEy", "avatar_url": "https://avatars.githubusercontent.com/u/70580312?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hansaskov", "html_url": "https://github.com/hansaskov", "followers_url": "https://api.github.com/users/hansaskov/followers", "following_url": "https://api.github.com/users/hansaskov/following{/other_user}", "gists_url": "https://api.github.com/users/hansaskov/gists{/gist_id}", "starred_url": "https://api.github.com/users/hansaskov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hansaskov/subscriptions", "organizations_url": "https://api.github.com/users/hansaskov/orgs", "repos_url": "https://api.github.com/users/hansaskov/repos", "events_url": "https://api.github.com/users/hansaskov/events{/privacy}", "received_events_url": "https://api.github.com/users/hansaskov/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-03-24T22:08:56
2024-03-24T22:17:04
2024-03-24T22:17:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am unable to download and install ollama on ubuntu 22.04. I used the following command ```bash sudo curl -fsSL https://ollama.com/install.sh | sh ``` And i get the following output ``` >>> Downloading ollama... Warning: Failed to open the file /tmp/tmp.uSBa9PbYEo/ollama: No such file or %##O#-# Warning: directory 0.0%curl: (23) Failure writing output to destination ``` Here is the output i get by just running `sudo curl -fsSL https://ollama.com/install.sh ` standalone ``` sh #!/bin/sh # This script installs Ollama on Linux. # It detects the current operating system architecture and installs the appropriate version of Ollama. set -eu status() { echo ">>> $*" >&2; } error() { echo "ERROR $*"; exit 1; } warning() { echo "WARNING: $*"; } TEMP_DIR=$(mktemp -d) cleanup() { rm -rf $TEMP_DIR; } trap cleanup EXIT available() { command -v $1 >/dev/null; } require() { local MISSING='' for TOOL in $*; do if ! available $TOOL; then MISSING="$MISSING $TOOL" fi done echo $MISSING } [ "$(uname -s)" = "Linux" ] || error 'This script is intended to run on Linux only.' ARCH=$(uname -m) case "$ARCH" in x86_64) ARCH="amd64" ;; aarch64|arm64) ARCH="arm64" ;; *) error "Unsupported architecture: $ARCH" ;; esac KERN=$(uname -r) case "$KERN" in *icrosoft*WSL2 | *icrosoft*wsl2) ;; *icrosoft) error "Microsoft WSL1 is not currently supported. Please upgrade to WSL2 with 'wsl --set-version <distro> 2'" ;; *) ;; esac VER_PARAM="${OLLAMA_VERSION:+?version=$OLLAMA_VERSION}" SUDO= if [ "$(id -u)" -ne 0 ]; then # Running as root, no need for sudo if ! available sudo; then error "This script requires superuser permissions. Please re-run as root." fi SUDO="sudo" fi NEEDS=$(require curl awk grep sed tee xargs) if [ -n "$NEEDS" ]; then status "ERROR: The following tools are required but missing:" for NEED in $NEEDS; do echo " - $NEED" done exit 1 fi status "Downloading ollama..." curl --fail --show-error --location --progress-bar -o $TEMP_DIR/ollama "https://ollama.com/download/ollama-linux-${ARCH}${VER_PARAM}" for BINDIR in /usr/local/bin /usr/bin /bin; do echo $PATH | grep -q $BINDIR && break || continue done status "Installing ollama to $BINDIR..." $SUDO install -o0 -g0 -m755 -d $BINDIR $SUDO install -o0 -g0 -m755 $TEMP_DIR/ollama $BINDIR/ollama install_success() { status 'The Ollama API is now available at 127.0.0.1:11434.' status 'Install complete. Run "ollama" from the command line.' } trap install_success EXIT # Everything from this point onwards is optional. configure_systemd() { if ! id ollama >/dev/null 2>&1; then status "Creating ollama user..." $SUDO useradd -r -s /bin/false -U -m -d /usr/share/ollama ollama fi if getent group render >/dev/null 2>&1; then status "Adding ollama user to render group..." $SUDO usermod -a -G render ollama fi if getent group video >/dev/null 2>&1; then status "Adding ollama user to video group..." $SUDO usermod -a -G video ollama fi status "Adding current user to ollama group..." $SUDO usermod -a -G ollama $(whoami) status "Creating ollama systemd service..." cat <<EOF | $SUDO tee /etc/systemd/system/ollama.service >/dev/null [Unit] Description=Ollama Service After=network-online.target [Service] ExecStart=$BINDIR/ollama serve User=ollama Group=ollama Restart=always RestartSec=3 Environment="PATH=$PATH" [Install] WantedBy=default.target EOF SYSTEMCTL_RUNNING="$(systemctl is-system-running || true)" case $SYSTEMCTL_RUNNING in running|degraded) status "Enabling and starting ollama service..." $SUDO systemctl daemon-reload $SUDO systemctl enable ollama start_service() { $SUDO systemctl restart ollama; } trap start_service EXIT ;; esac } if available systemctl; then configure_systemd fi if ! available lspci && ! available lshw; then warning "Unable to detect NVIDIA/AMD GPU. Install lspci or lshw to automatically detect and install GPU dependencies." exit 0 fi check_gpu() { # Look for devices based on vendor ID for NVIDIA and AMD case $1 in lspci) case $2 in nvidia) available lspci && lspci -d '10de:' | grep -q 'NVIDIA' || return 1 ;; amdgpu) available lspci && lspci -d '1002:' | grep -q 'AMD' || return 1 ;; esac ;; lshw) case $2 in nvidia) available lshw && $SUDO lshw -c display -numeric | grep -q 'vendor: .* \[10DE\]' || return 1 ;; amdgpu) available lshw && $SUDO lshw -c display -numeric | grep -q 'vendor: .* \[1002\]' || return 1 ;; esac ;; nvidia-smi) available nvidia-smi || return 1 ;; esac } if check_gpu nvidia-smi; then status "NVIDIA GPU installed." exit 0 fi if ! check_gpu lspci nvidia && ! check_gpu lshw nvidia && ! check_gpu lspci amdgpu && ! check_gpu lshw amdgpu; then install_success warning "No NVIDIA/AMD GPU detected. Ollama will run in CPU-only mode." exit 0 fi if check_gpu lspci amdgpu || check_gpu lshw amdgpu; then # Look for pre-existing ROCm v6 before downloading the dependencies for search in "${HIP_PATH:-''}" "${ROCM_PATH:-''}" "/opt/rocm"; do if [ -n "${search}" ] && [ -e "${search}/lib/libhipblas.so.2" ]; then status "Compatible AMD GPU ROCm library detected at ${search}" install_success exit 0 fi done status "Downloading AMD GPU dependencies..." $SUDO rm -rf /usr/share/ollama/lib $SUDO chmod o+x /usr/share/ollama $SUDO install -o ollama -g ollama -m 755 -d /usr/share/ollama/lib/rocm curl --fail --show-error --location --progress-bar "https://ollama.com/download/ollama-linux-amd64-rocm.tgz${VER_PARAM}" \ | $SUDO tar zx --owner ollama --group ollama -C /usr/share/ollama/lib/rocm . install_success status "AMD GPU dependencies installed." exit 0 fi # ref: https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html#rhel-7-centos-7 # ref: https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html#rhel-8-rocky-8 # ref: https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html#rhel-9-rocky-9 # ref: https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html#fedora install_cuda_driver_yum() { status 'Installing NVIDIA repository...' case $PACKAGE_MANAGER in yum) $SUDO $PACKAGE_MANAGER -y install yum-utils $SUDO $PACKAGE_MANAGER-config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/$1$2/$(uname -m)/cuda-$1$2.repo ;; dnf) $SUDO $PACKAGE_MANAGER config-manager --add-repo https://developer.download.nvidia.com/compute/cuda/repos/$1$2/$(uname -m)/cuda-$1$2.repo ;; esac case $1 in rhel) status 'Installing EPEL repository...' # EPEL is required for third-party dependencies such as dkms and libvdpau $SUDO $PACKAGE_MANAGER -y install https://dl.fedoraproject.org/pub/epel/epel-release-latest-$2.noarch.rpm || true ;; esac status 'Installing CUDA driver...' if [ "$1" = 'centos' ] || [ "$1$2" = 'rhel7' ]; then $SUDO $PACKAGE_MANAGER -y install nvidia-driver-latest-dkms fi $SUDO $PACKAGE_MANAGER -y install cuda-drivers } # ref: https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html#ubuntu # ref: https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html#debian install_cuda_driver_apt() { status 'Installing NVIDIA repository...' curl -fsSL -o $TEMP_DIR/cuda-keyring.deb https://developer.download.nvidia.com/compute/cuda/repos/$1$2/$(uname -m)/cuda-keyring_1.1-1_all.deb case $1 in debian) status 'Enabling contrib sources...' $SUDO sed 's/main/contrib/' < /etc/apt/sources.list | $SUDO tee /etc/apt/sources.list.d/contrib.list > /dev/null if [ -f "/etc/apt/sources.list.d/debian.sources" ]; then $SUDO sed 's/main/contrib/' < /etc/apt/sources.list.d/debian.sources | $SUDO tee /etc/apt/sources.list.d/contrib.sources > /dev/null fi ;; esac status 'Installing CUDA driver...' $SUDO dpkg -i $TEMP_DIR/cuda-keyring.deb $SUDO apt-get update [ -n "$SUDO" ] && SUDO_E="$SUDO -E" || SUDO_E= DEBIAN_FRONTEND=noninteractive $SUDO_E apt-get -y install cuda-drivers -q } if [ ! -f "/etc/os-release" ]; then error "Unknown distribution. Skipping CUDA installation." fi . /etc/os-release OS_NAME=$ID OS_VERSION=$VERSION_ID PACKAGE_MANAGER= for PACKAGE_MANAGER in dnf yum apt-get; do if available $PACKAGE_MANAGER; then break fi done if [ -z "$PACKAGE_MANAGER" ]; then error "Unknown package manager. Skipping CUDA installation." fi if ! check_gpu nvidia-smi || [ -z "$(nvidia-smi | grep -o "CUDA Version: [0-9]*\.[0-9]*")" ]; then case $OS_NAME in centos|rhel) install_cuda_driver_yum 'rhel' $(echo $OS_VERSION | cut -d '.' -f 1) ;; rocky) install_cuda_driver_yum 'rhel' $(echo $OS_VERSION | cut -c1) ;; fedora) [ $OS_VERSION -lt '37' ] && install_cuda_driver_yum $OS_NAME $OS_VERSION || install_cuda_driver_yum $OS_NAME '37';; amzn) install_cuda_driver_yum 'fedora' '37' ;; debian) install_cuda_driver_apt $OS_NAME $OS_VERSION ;; ubuntu) install_cuda_driver_apt $OS_NAME $(echo $OS_VERSION | sed 's/\.//') ;; *) exit ;; esac fi if ! lsmod | grep -q nvidia; then KERNEL_RELEASE="$(uname -r)" case $OS_NAME in rocky) $SUDO $PACKAGE_MANAGER -y install kernel-devel kernel-headers ;; centos|rhel|amzn) $SUDO $PACKAGE_MANAGER -y install kernel-devel-$KERNEL_RELEASE kernel-headers-$KERNEL_RELEASE ;; fedora) $SUDO $PACKAGE_MANAGER -y install kernel-devel-$KERNEL_RELEASE ;; debian|ubuntu) $SUDO apt-get -y install linux-headers-$KERNEL_RELEASE ;; *) exit ;; esac NVIDIA_CUDA_VERSION=$($SUDO dkms status | awk -F: '/added/ { print $1 }') if [ -n "$NVIDIA_CUDA_VERSION" ]; then $SUDO dkms install $NVIDIA_CUDA_VERSION fi if lsmod | grep -q nouveau; then status 'Reboot to complete NVIDIA CUDA driver install.' exit 0 fi $SUDO modprobe nvidia fi status "NVIDIA CUDA drivers installed." ``` ### What did you expect to see? I expected the ollama to be installed ### Steps to reproduce Run `sudo curl -fsSL https://ollama.com/install.sh | sh` ### Are there any recent changes that introduced the issue? I would not say so ### OS Linux ### Architecture amd64 ### Platform _No response_ ### Ollama version _No response_ ### GPU Nvidia, AMD ### GPU info GTX 1080 ti AMD R5 7600 integrated graphics nvidia-smi ``` nvidia-smi Sun Mar 24 23:09:06 2024 +-----------------------------------------------------------------------------+ | NVIDIA-SMI 470.239.06 Driver Version: 470.239.06 CUDA Version: 11.4 | |-------------------------------+----------------------+----------------------+ | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |===============================+======================+======================| | 0 NVIDIA GeForce ... Off | 00000000:01:00.0 On | N/A | | 28% 32C P2 52W / 250W | 109MiB / 11178MiB | 16% Default | | | | N/A | +-------------------------------+----------------------+----------------------+ +-----------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=============================================================================| | 0 N/A N/A 1189 G /usr/lib/xorg/Xorg 107MiB | +-----------------------------------------------------------------------------+ ``` ### CPU AMD R5 7600 ### Other software
{ "login": "hansaskov", "id": 70580312, "node_id": "MDQ6VXNlcjcwNTgwMzEy", "avatar_url": "https://avatars.githubusercontent.com/u/70580312?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hansaskov", "html_url": "https://github.com/hansaskov", "followers_url": "https://api.github.com/users/hansaskov/followers", "following_url": "https://api.github.com/users/hansaskov/following{/other_user}", "gists_url": "https://api.github.com/users/hansaskov/gists{/gist_id}", "starred_url": "https://api.github.com/users/hansaskov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hansaskov/subscriptions", "organizations_url": "https://api.github.com/users/hansaskov/orgs", "repos_url": "https://api.github.com/users/hansaskov/repos", "events_url": "https://api.github.com/users/hansaskov/events{/privacy}", "received_events_url": "https://api.github.com/users/hansaskov/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3330/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3330/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8271
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8271/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8271/comments
https://api.github.com/repos/ollama/ollama/issues/8271/events
https://github.com/ollama/ollama/issues/8271
2,763,974,531
I_kwDOJ0Z1Ps6kvueD
8,271
llama runner process terminated: CUDA error
{ "login": "iplayfast", "id": 751306, "node_id": "MDQ6VXNlcjc1MTMwNg==", "avatar_url": "https://avatars.githubusercontent.com/u/751306?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iplayfast", "html_url": "https://github.com/iplayfast", "followers_url": "https://api.github.com/users/iplayfast/followers", "following_url": "https://api.github.com/users/iplayfast/following{/other_user}", "gists_url": "https://api.github.com/users/iplayfast/gists{/gist_id}", "starred_url": "https://api.github.com/users/iplayfast/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iplayfast/subscriptions", "organizations_url": "https://api.github.com/users/iplayfast/orgs", "repos_url": "https://api.github.com/users/iplayfast/repos", "events_url": "https://api.github.com/users/iplayfast/events{/privacy}", "received_events_url": "https://api.github.com/users/iplayfast/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
14
2024-12-31T05:34:54
2025-01-01T07:33:02
2025-01-01T07:33:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm making a little embedding example. The text is chunked fine, and embedded. But on retrieval, the I get this weird message. Then if I go to the cli and try running the same model, I get the same message. I'm pretty sure it has to do with having the embedded model in memory at the same time as llama3.2 **Also at the time only the embedded model was in memory, llama3.2 was supposed to be (and had just been used) but is no longer there, which I think is a big clue** ``` (venv-game) chris@FORGE:~/game$ ollama run llama3.2:latest Error: llama runner process has terminated: CUDA error (venv-game) chris@FORGE:~/game$ ollama ps NAME ID SIZE PROCESSOR UNTIL nomic-embed-text:latest 0a109f422b47 849 MB 100% GPU About a minute from now (venv-game) chris@FORGE:~/game$ ollama run llama3.2:latest Error: llama runner process has terminated: CUDA error (venv-game) chris@FORGE:~/game$ nvidia-smi Tue Dec 31 00:33:00 2024 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.120 Driver Version: 550.120 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce RTX 4090 On | 00000000:01:00.0 On | Off | | 0% 42C P8 23W / 450W | 2269MiB / 24564MiB | 16% E. Process | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | 0 N/A N/A 2454 G /usr/lib/xorg/Xorg 653MiB | | 0 N/A N/A 3508 G cinnamon 71MiB | | 0 N/A N/A 7665 G ...seed-version=20241225-174432.450000 341MiB | | 0 N/A N/A 592670 G ...erProcess --variations-seed-version 76MiB | | 0 N/A N/A 1334610 C ...rs/cuda_v12_avx/ollama_llama_server 930MiB | | 0 N/A N/A 1663080 G ...yOnDemand --variations-seed-version 105MiB | | 0 N/A N/A 3924325 G ...erProcess --variations-seed-version 61MiB | +-----------------------------------------------------------------------------------------+ ``` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version ollama version is 0.5.4
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8271/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8271/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7839
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7839/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7839/comments
https://api.github.com/repos/ollama/ollama/issues/7839/events
https://github.com/ollama/ollama/pull/7839
2,693,432,609
PR_kwDOJ0Z1Ps6DI6lo
7,839
fix(docs): Replace hard coded num of make threads w/ CPU cores
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/followers", "following_url": "https://api.github.com/users/sammcj/following{/other_user}", "gists_url": "https://api.github.com/users/sammcj/gists{/gist_id}", "starred_url": "https://api.github.com/users/sammcj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sammcj/subscriptions", "organizations_url": "https://api.github.com/users/sammcj/orgs", "repos_url": "https://api.github.com/users/sammcj/repos", "events_url": "https://api.github.com/users/sammcj/events{/privacy}", "received_events_url": "https://api.github.com/users/sammcj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-11-26T06:54:28
2024-12-06T22:51:09
2024-12-06T22:51:04
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7839", "html_url": "https://github.com/ollama/ollama/pull/7839", "diff_url": "https://github.com/ollama/ollama/pull/7839.diff", "patch_url": "https://github.com/ollama/ollama/pull/7839.patch", "merged_at": null }
Replace the seemingly random number of make threads with the number CPU cores / 2.
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/followers", "following_url": "https://api.github.com/users/sammcj/following{/other_user}", "gists_url": "https://api.github.com/users/sammcj/gists{/gist_id}", "starred_url": "https://api.github.com/users/sammcj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sammcj/subscriptions", "organizations_url": "https://api.github.com/users/sammcj/orgs", "repos_url": "https://api.github.com/users/sammcj/repos", "events_url": "https://api.github.com/users/sammcj/events{/privacy}", "received_events_url": "https://api.github.com/users/sammcj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7839/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7839/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3914
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3914/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3914/comments
https://api.github.com/repos/ollama/ollama/issues/3914/events
https://github.com/ollama/ollama/pull/3914
2,264,038,325
PR_kwDOJ0Z1Ps5twARO
3,914
Improve mac parallel performance
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-25T16:40:14
2024-04-25T23:28:34
2024-04-25T23:28:31
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3914", "html_url": "https://github.com/ollama/ollama/pull/3914", "diff_url": "https://github.com/ollama/ollama/pull/3914.diff", "patch_url": "https://github.com/ollama/ollama/pull/3914.patch", "merged_at": "2024-04-25T23:28:31" }
Carries #3900 Testing underway...
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3914/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3914/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4048
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4048/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4048/comments
https://api.github.com/repos/ollama/ollama/issues/4048/events
https://github.com/ollama/ollama/issues/4048
2,271,291,702
I_kwDOJ0Z1Ps6HYSk2
4,048
Windows installation is behaving like malware
{ "login": "eabase", "id": 52289379, "node_id": "MDQ6VXNlcjUyMjg5Mzc5", "avatar_url": "https://avatars.githubusercontent.com/u/52289379?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eabase", "html_url": "https://github.com/eabase", "followers_url": "https://api.github.com/users/eabase/followers", "following_url": "https://api.github.com/users/eabase/following{/other_user}", "gists_url": "https://api.github.com/users/eabase/gists{/gist_id}", "starred_url": "https://api.github.com/users/eabase/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eabase/subscriptions", "organizations_url": "https://api.github.com/users/eabase/orgs", "repos_url": "https://api.github.com/users/eabase/repos", "events_url": "https://api.github.com/users/eabase/events{/privacy}", "received_events_url": "https://api.github.com/users/eabase/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
8
2024-04-30T12:06:32
2024-05-08T12:43:37
2024-05-01T16:34:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? This is beyond bad. 1. It installs stuff all over the place, without informing the user. ``` C:\Users\xxxx\AppData\Roaming\Microsoft\Windows\Start Menu\Programs C:\Users\xxxx\AppData\Local\Temp\ollamaNNNNNN C:\Users\xxxx\.ollama\id_ed25519 C:\Users\xxxx\AppData\Local\Programs\Ollama C:\Users\xxxx\AppData\Local\Ollama ``` 2. It doesn't ask the user where they want it to be installed! 3. It opens ports and starts a server under different application name and on unspecified sever ports that can only be found in a temp directory. 4. Once the original (cmd) windows is closed, the server is still running! 5. There are no instructions on how to shut down the server In the current state this application is acting more like **malware** than something serious. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version _No response_
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4048/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4048/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3025
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3025/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3025/comments
https://api.github.com/repos/ollama/ollama/issues/3025/events
https://github.com/ollama/ollama/pull/3025
2,177,367,153
PR_kwDOJ0Z1Ps5pJu-X
3,025
Rename ROCm deps file to avoid confusion
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-09T19:22:33
2024-03-10T03:50:45
2024-03-10T01:48:39
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3025", "html_url": "https://github.com/ollama/ollama/pull/3025", "diff_url": "https://github.com/ollama/ollama/pull/3025.diff", "patch_url": "https://github.com/ollama/ollama/pull/3025.patch", "merged_at": "2024-03-10T01:48:39" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3025/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3025/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7111
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7111/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7111/comments
https://api.github.com/repos/ollama/ollama/issues/7111/events
https://github.com/ollama/ollama/issues/7111
2,569,542,877
I_kwDOJ0Z1Ps6ZKBzd
7,111
Any plans to add nvidia/NVLM-D-72B new model to ollama?
{ "login": "SilverLaius", "id": 26542231, "node_id": "MDQ6VXNlcjI2NTQyMjMx", "avatar_url": "https://avatars.githubusercontent.com/u/26542231?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SilverLaius", "html_url": "https://github.com/SilverLaius", "followers_url": "https://api.github.com/users/SilverLaius/followers", "following_url": "https://api.github.com/users/SilverLaius/following{/other_user}", "gists_url": "https://api.github.com/users/SilverLaius/gists{/gist_id}", "starred_url": "https://api.github.com/users/SilverLaius/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SilverLaius/subscriptions", "organizations_url": "https://api.github.com/users/SilverLaius/orgs", "repos_url": "https://api.github.com/users/SilverLaius/repos", "events_url": "https://api.github.com/users/SilverLaius/events{/privacy}", "received_events_url": "https://api.github.com/users/SilverLaius/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
1
2024-10-07T07:08:00
2024-10-07T07:09:13
2024-10-07T07:08:51
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Nvidia just released a state of the art model that would be a good addition to ollama. Any plans to add it? [https://huggingface.co/nvidia/NVLM-D-72B/tree/main](https://huggingface.co/nvidia/NVLM-D-72B/tree/main)
{ "login": "SilverLaius", "id": 26542231, "node_id": "MDQ6VXNlcjI2NTQyMjMx", "avatar_url": "https://avatars.githubusercontent.com/u/26542231?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SilverLaius", "html_url": "https://github.com/SilverLaius", "followers_url": "https://api.github.com/users/SilverLaius/followers", "following_url": "https://api.github.com/users/SilverLaius/following{/other_user}", "gists_url": "https://api.github.com/users/SilverLaius/gists{/gist_id}", "starred_url": "https://api.github.com/users/SilverLaius/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SilverLaius/subscriptions", "organizations_url": "https://api.github.com/users/SilverLaius/orgs", "repos_url": "https://api.github.com/users/SilverLaius/repos", "events_url": "https://api.github.com/users/SilverLaius/events{/privacy}", "received_events_url": "https://api.github.com/users/SilverLaius/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7111/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7111/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7619
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7619/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7619/comments
https://api.github.com/repos/ollama/ollama/issues/7619/events
https://github.com/ollama/ollama/issues/7619
2,648,701,609
I_kwDOJ0Z1Ps6d3_qp
7,619
llama3.2-vision on multi gpu error
{ "login": "18600709862", "id": 46145698, "node_id": "MDQ6VXNlcjQ2MTQ1Njk4", "avatar_url": "https://avatars.githubusercontent.com/u/46145698?v=4", "gravatar_id": "", "url": "https://api.github.com/users/18600709862", "html_url": "https://github.com/18600709862", "followers_url": "https://api.github.com/users/18600709862/followers", "following_url": "https://api.github.com/users/18600709862/following{/other_user}", "gists_url": "https://api.github.com/users/18600709862/gists{/gist_id}", "starred_url": "https://api.github.com/users/18600709862/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/18600709862/subscriptions", "organizations_url": "https://api.github.com/users/18600709862/orgs", "repos_url": "https://api.github.com/users/18600709862/repos", "events_url": "https://api.github.com/users/18600709862/events{/privacy}", "received_events_url": "https://api.github.com/users/18600709862/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
null
[]
null
7
2024-11-11T09:26:43
2024-11-12T20:59:09
2024-11-12T20:58:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? multi gpu ollama run llama3.2-vision >>> The image is a book cover. Output should be in this format - <Name of the Book>: <Name of the Author>. Do not output anything else /media/root/ssd2t/data/pro/tmp/o ... l/new/FastChat/image.png Added image '/media/root/ssd2t/data/pro/tmp/ol/new/FastChat/image.png' Error: POST predict: Post "http://127.0.0.1:41121/completion": EOF one gpu run ok ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version 0.4.1
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7619/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7619/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/748
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/748/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/748/comments
https://api.github.com/repos/ollama/ollama/issues/748/events
https://github.com/ollama/ollama/issues/748
1,935,466,908
I_kwDOJ0Z1Ps5zXOGc
748
Enabling langchain
{ "login": "suoko", "id": 3659980, "node_id": "MDQ6VXNlcjM2NTk5ODA=", "avatar_url": "https://avatars.githubusercontent.com/u/3659980?v=4", "gravatar_id": "", "url": "https://api.github.com/users/suoko", "html_url": "https://github.com/suoko", "followers_url": "https://api.github.com/users/suoko/followers", "following_url": "https://api.github.com/users/suoko/following{/other_user}", "gists_url": "https://api.github.com/users/suoko/gists{/gist_id}", "starred_url": "https://api.github.com/users/suoko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/suoko/subscriptions", "organizations_url": "https://api.github.com/users/suoko/orgs", "repos_url": "https://api.github.com/users/suoko/repos", "events_url": "https://api.github.com/users/suoko/events{/privacy}", "received_events_url": "https://api.github.com/users/suoko/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-10-10T14:13:06
2023-10-11T15:45:32
2023-10-11T15:45:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I found this video about implementing langchain with ollama. If available it should be easy to use it via ollama-ui https://www.youtube.com/watch?v=CPgp8MhmGVY
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/748/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/748/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2844
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2844/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2844/comments
https://api.github.com/repos/ollama/ollama/issues/2844/events
https://github.com/ollama/ollama/issues/2844
2,162,130,461
I_kwDOJ0Z1Ps6A334d
2,844
OpenAI package compatibility
{ "login": "eliranwong", "id": 25262722, "node_id": "MDQ6VXNlcjI1MjYyNzIy", "avatar_url": "https://avatars.githubusercontent.com/u/25262722?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eliranwong", "html_url": "https://github.com/eliranwong", "followers_url": "https://api.github.com/users/eliranwong/followers", "following_url": "https://api.github.com/users/eliranwong/following{/other_user}", "gists_url": "https://api.github.com/users/eliranwong/gists{/gist_id}", "starred_url": "https://api.github.com/users/eliranwong/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eliranwong/subscriptions", "organizations_url": "https://api.github.com/users/eliranwong/orgs", "repos_url": "https://api.github.com/users/eliranwong/repos", "events_url": "https://api.github.com/users/eliranwong/events{/privacy}", "received_events_url": "https://api.github.com/users/eliranwong/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
6
2024-02-29T21:26:43
2024-05-02T01:23:23
2024-03-01T01:08:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I read the example in https://ollama.com/blog/openai-compatibility The example works, but it doesn't when I add "response_format={ "type": "json_object" }," https://platform.openai.com/docs/guides/text-generation/json-mode
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2844/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2844/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/285
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/285/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/285/comments
https://api.github.com/repos/ollama/ollama/issues/285/events
https://github.com/ollama/ollama/pull/285
1,837,254,131
PR_kwDOJ0Z1Ps5XOXrT
285
Update README.md
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-08-04T19:43:08
2023-08-04T19:45:56
2023-08-04T19:45:55
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/285", "html_url": "https://github.com/ollama/ollama/pull/285", "diff_url": "https://github.com/ollama/ollama/pull/285.diff", "patch_url": "https://github.com/ollama/ollama/pull/285.patch", "merged_at": "2023-08-04T19:45:55" }
Ollama now supports Intel Macs
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/285/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/285/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2392
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2392/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2392/comments
https://api.github.com/repos/ollama/ollama/issues/2392/events
https://github.com/ollama/ollama/issues/2392
2,123,591,240
I_kwDOJ0Z1Ps5-k25I
2,392
unable to initialize llm library Radeon card detected
{ "login": "sigmaya", "id": 159318222, "node_id": "U_kgDOCX8Azg", "avatar_url": "https://avatars.githubusercontent.com/u/159318222?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sigmaya", "html_url": "https://github.com/sigmaya", "followers_url": "https://api.github.com/users/sigmaya/followers", "following_url": "https://api.github.com/users/sigmaya/following{/other_user}", "gists_url": "https://api.github.com/users/sigmaya/gists{/gist_id}", "starred_url": "https://api.github.com/users/sigmaya/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sigmaya/subscriptions", "organizations_url": "https://api.github.com/users/sigmaya/orgs", "repos_url": "https://api.github.com/users/sigmaya/repos", "events_url": "https://api.github.com/users/sigmaya/events{/privacy}", "received_events_url": "https://api.github.com/users/sigmaya/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
6
2024-02-07T18:04:01
2024-02-29T00:50:07
2024-02-07T22:47:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello, I am trying to run as user and manually, i get this error: time=2024-02-07T19:00:18.967+01:00 level=INFO source=payload_common.go:106 msg="Extracting dynamic libraries..." Error: unable to initialize llm library Radeon card detected, but permissions not set up properly. Either run ollama as root, or add you user account to the render group. I had a firepro w7100 but some days ago , i removed it and now i am using an nvidia 3060, I am on ubuntu 20 and i have no idea how to tell ollama that the gpu is nvidia.
{ "login": "sigmaya", "id": 159318222, "node_id": "U_kgDOCX8Azg", "avatar_url": "https://avatars.githubusercontent.com/u/159318222?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sigmaya", "html_url": "https://github.com/sigmaya", "followers_url": "https://api.github.com/users/sigmaya/followers", "following_url": "https://api.github.com/users/sigmaya/following{/other_user}", "gists_url": "https://api.github.com/users/sigmaya/gists{/gist_id}", "starred_url": "https://api.github.com/users/sigmaya/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sigmaya/subscriptions", "organizations_url": "https://api.github.com/users/sigmaya/orgs", "repos_url": "https://api.github.com/users/sigmaya/repos", "events_url": "https://api.github.com/users/sigmaya/events{/privacy}", "received_events_url": "https://api.github.com/users/sigmaya/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2392/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2392/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5488
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5488/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5488/comments
https://api.github.com/repos/ollama/ollama/issues/5488/events
https://github.com/ollama/ollama/issues/5488
2,391,347,345
I_kwDOJ0Z1Ps6OiRCR
5,488
falcon2 fails to genarate code in almalinux9
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/followers", "following_url": "https://api.github.com/users/olumolu/following{/other_user}", "gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}", "starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/olumolu/subscriptions", "organizations_url": "https://api.github.com/users/olumolu/orgs", "repos_url": "https://api.github.com/users/olumolu/repos", "events_url": "https://api.github.com/users/olumolu/events{/privacy}", "received_events_url": "https://api.github.com/users/olumolu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
0
2024-07-04T18:55:46
2024-09-26T12:39:47
2024-09-26T12:39:47
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I got a error that it could not able to genarate any code but it failed and just skipped and jumped 5lines of free space and ask for input again. ### OS Linux ### GPU _No response_ ### CPU Intel ### Ollama version 1.48
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/followers", "following_url": "https://api.github.com/users/olumolu/following{/other_user}", "gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}", "starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/olumolu/subscriptions", "organizations_url": "https://api.github.com/users/olumolu/orgs", "repos_url": "https://api.github.com/users/olumolu/repos", "events_url": "https://api.github.com/users/olumolu/events{/privacy}", "received_events_url": "https://api.github.com/users/olumolu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5488/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5488/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3287
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3287/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3287/comments
https://api.github.com/repos/ollama/ollama/issues/3287/events
https://github.com/ollama/ollama/issues/3287
2,200,586,023
I_kwDOJ0Z1Ps6DKkcn
3,287
Allow system message to be on its own
{ "login": "jackielii", "id": 360983, "node_id": "MDQ6VXNlcjM2MDk4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jackielii", "html_url": "https://github.com/jackielii", "followers_url": "https://api.github.com/users/jackielii/followers", "following_url": "https://api.github.com/users/jackielii/following{/other_user}", "gists_url": "https://api.github.com/users/jackielii/gists{/gist_id}", "starred_url": "https://api.github.com/users/jackielii/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jackielii/subscriptions", "organizations_url": "https://api.github.com/users/jackielii/orgs", "repos_url": "https://api.github.com/users/jackielii/repos", "events_url": "https://api.github.com/users/jackielii/events{/privacy}", "received_events_url": "https://api.github.com/users/jackielii/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-03-21T16:05:18
2024-09-04T04:42:37
2024-09-04T04:42:37
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? Current prompt rendering assumes there is only one system message. With the [Command-R model from Cohere AI](https://huggingface.co/CohereForAI/c4ai-command-r-v01), there can be multiple system message: https://docs.cohere.com/docs/prompting-command-r#structured-prompt-templates. There is no way to get around this. Consider the following messages: ```go {Role: "system", Content: "Preambles"}, {Role: "user", Content: "i'm user"}, {Role: "system", Content: "Tool output"}, {Role: "system", Content: "---Instructions---"}, ``` And the template: ``` <BOS_TOKEN> {{ if .System }}<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{{ .System }}<|END_OF_TURN_TOKEN|>{{ end }} {{ if .Prompt }}<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{{ .Prompt }}<|END_OF_TURN_TOKEN|>{{ end }}<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>{{ .Response }} ``` We get ``` <BOS_TOKEN> <|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>Preambles<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|USER_TOKEN|>i'm user<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><BOS_TOKEN> <|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>Tool output<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><BOS_TOKEN> <|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>---Instructions---<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|> ``` However, the desired output is: ``` <BOS_TOKEN> <|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>Preambles<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|USER_TOKEN|>i'm user<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><BOS_TOKEN> <|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>Tool output<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>---Instructions---<|END_OF_TURN_TOKEN|> <|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|> ``` ### How should we solve this? In order to make this backward compatible, I suggest we introduce a new version of the template that formats each message as is. It'll be much easier to control the desired output The code to render the prompt should detect is the new template is present and use it instead. Preferably, gradually deprecate old template. ### What is the impact of not solving this? There doesn't seems to be a backward compatible way to fix this. Introducing a new template add complexity, especially for new users. But there doesn't seems to be a better way. ### Anything else? _No response_
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3287/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3287/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3711
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3711/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3711/comments
https://api.github.com/repos/ollama/ollama/issues/3711/events
https://github.com/ollama/ollama/issues/3711
2,249,202,461
I_kwDOJ0Z1Ps6GEBsd
3,711
CUDA malloc fails on newly supported models in 0.1.32 (dual-GPU setup with 72GB VRAM and 128GB RAM)
{ "login": "mz2", "id": 71363, "node_id": "MDQ6VXNlcjcxMzYz", "avatar_url": "https://avatars.githubusercontent.com/u/71363?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mz2", "html_url": "https://github.com/mz2", "followers_url": "https://api.github.com/users/mz2/followers", "following_url": "https://api.github.com/users/mz2/following{/other_user}", "gists_url": "https://api.github.com/users/mz2/gists{/gist_id}", "starred_url": "https://api.github.com/users/mz2/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mz2/subscriptions", "organizations_url": "https://api.github.com/users/mz2/orgs", "repos_url": "https://api.github.com/users/mz2/repos", "events_url": "https://api.github.com/users/mz2/events{/privacy}", "received_events_url": "https://api.github.com/users/mz2/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
16
2024-04-17T20:18:35
2024-06-22T00:10:31
2024-06-22T00:10:31
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am getting cuda malloc errors with v0.1.32 (as well as with the current head of main branch) when trying any of the new big models: wizardlm2, mixtral:8x22b, dbrx (command-r+ does work) with my dual GPU setup (A6000 + RTX 3090, i.e. combined 72GB VRAM) with a 24-core 13th gen Intel CPU (128GB of DDR5 on the system). The symptoms are similar, with llama-cpp dying ``` ollama run mixtral:8x22b Error: llama runner process no longer running: 1 error:failed to create context with model '/media/data/ollama/blobs/sha256-b5fc1eb35edf792b07d6163cf7ac162fdd9f9024903e6b33a3a870f2f973b8ca' ``` From the service logs I see: ``` Apr 17 23:09:42 athena ollama.listener[2840147]: time=2024-04-17T23:09:42.621+03:00 level=INFO source=server.go:136 msg="offload to gpu" layers.real=39 layers.estimate=39 memory.available="70186.6 MiB" memory.required.full="72169.5 MiB" memory.required.partial="69599.9 MiB" memory.required.kv="320.0 MiB" memory.weights.total="70752.5 MiB" memory.weights.repeating="69939.4 MiB" memory.weights.nonrepeating="813.1 MiB" memory.graph.full="640.0 MiB" memory.graph.partial="640.0 MiB" Apr 17 23:09:42 athena ollama.listener[2840147]: time=2024-04-17T23:09:42.621+03:00 level=INFO source=cpu_common.go:11 msg="CPU has AVX2" Apr 17 23:09:42 athena ollama.listener[2840147]: time=2024-04-17T23:09:42.622+03:00 level=INFO source=server.go:302 msg="starting llama server" cmd="/tmp/ollama3763278260/runners/cuda_v12/ollama_llama_server --model /media/data/ollama/blobs/sha256-1d12441f19436dbb0bcc4067e9d47921b944ef4a87b35873aa430e85e91a93c8 --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 39 --port 38939" Apr 17 23:09:42 athena ollama.listener[2840147]: time=2024-04-17T23:09:42.622+03:00 level=INFO source=server.go:427 msg="waiting for llama runner to start responding" Apr 17 23:09:42 athena ollama.listener[2841851]: {"function":"server_params_parse","level":"INFO","line":2603,"msg":"logging to file is disabled.","tid":"123385345642496","timestamp":1713384582} Apr 17 23:09:42 athena ollama.listener[2841851]: {"build":1,"commit":"7593639","function":"main","level":"INFO","line":2819,"msg":"build info","tid":"123385345642496","timestamp":1713384582} Apr 17 23:09:42 athena ollama.listener[2841851]: {"function":"main","level":"INFO","line":2822,"msg":"system info","n_threads":16,"n_threads_batch":-1,"system_info":"AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | ","tid":"123385345642496","timestamp":1713384582,"total_threads":32} Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: loaded meta data with 24 key-value pairs and 323 tensors from /media/data/ollama/blobs/sha256-1d12441f19436dbb0bcc4067e9d47921b944ef4a87b35873aa430e85e91a93c8 (version GGUF V3 (latest)) Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 0: general.architecture str = dbrx Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 1: general.name str = dbrx Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 2: dbrx.block_count u32 = 40 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 3: dbrx.context_length u32 = 32768 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 4: dbrx.embedding_length u32 = 6144 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 5: dbrx.feed_forward_length u32 = 10752 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 6: dbrx.attention.head_count u32 = 48 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 7: dbrx.attention.head_count_kv u32 = 8 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 8: dbrx.rope.freq_base f32 = 500000.000000 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 9: dbrx.attention.clamp_kqv f32 = 8.000000 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 10: general.file_type u32 = 2 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 11: dbrx.expert_count u32 = 16 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 12: dbrx.expert_used_count u32 = 4 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 13: dbrx.attention.layer_norm_epsilon f32 = 0.000010 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 14: tokenizer.ggml.model str = gpt2 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,100352] = ["!", "\"", "#", "$", "%", "&", "'", ... Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 16: tokenizer.ggml.token_type arr[i32,100352] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 17: tokenizer.ggml.merges arr[str,100000] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 100257 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 19: tokenizer.ggml.eos_token_id u32 = 100257 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 20: tokenizer.ggml.unknown_token_id u32 = 100257 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 100277 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 22: tokenizer.chat_template str = {% if messages[0]['role'] == 'system'... Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - kv 23: general.quantization_version u32 = 2 Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - type f32: 81 tensors Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - type f16: 40 tensors Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - type q4_0: 201 tensors Apr 17 23:09:42 athena ollama.listener[2840147]: llama_model_loader: - type q6_K: 1 tensors Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_vocab: special tokens definition check successful ( 96/100352 ). Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: format = GGUF V3 (latest) Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: arch = dbrx Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: vocab type = BPE Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_vocab = 100352 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_merges = 100000 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_ctx_train = 32768 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_embd = 6144 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_head = 48 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_head_kv = 8 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_layer = 40 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_rot = 128 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_embd_head_k = 128 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_embd_head_v = 128 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_gqa = 6 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_embd_k_gqa = 1024 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_embd_v_gqa = 1024 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: f_norm_eps = 1.0e-05 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: f_norm_rms_eps = 0.0e+00 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: f_clamp_kqv = 8.0e+00 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: f_max_alibi_bias = 0.0e+00 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: f_logit_scale = 0.0e+00 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_ff = 10752 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_expert = 16 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_expert_used = 4 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: causal attn = 1 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: pooling type = 0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: rope type = 2 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: rope scaling = linear Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: freq_base_train = 500000.0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: freq_scale_train = 1 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: n_yarn_orig_ctx = 32768 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: rope_finetuned = unknown Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: ssm_d_conv = 0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: ssm_d_inner = 0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: ssm_d_state = 0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: ssm_dt_rank = 0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: model type = 16x12B Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: model ftype = Q4_0 Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: model params = 131.60 B Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: model size = 69.09 GiB (4.51 BPW) Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: general.name = dbrx Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: BOS token = 100257 '<|endoftext|>' Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: EOS token = 100257 '<|endoftext|>' Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: UNK token = 100257 '<|endoftext|>' Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: PAD token = 100277 '<|pad|>' Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_print_meta: LF token = 128 'Ä' Apr 17 23:09:42 athena ollama.listener[2840147]: ggml_cuda_init: GGML_CUDA_FORCE_MMQ: yes Apr 17 23:09:42 athena ollama.listener[2840147]: ggml_cuda_init: CUDA_USE_TENSOR_CORES: no Apr 17 23:09:42 athena ollama.listener[2840147]: ggml_cuda_init: found 2 CUDA devices: Apr 17 23:09:42 athena ollama.listener[2840147]: Device 0: NVIDIA RTX A6000, compute capability 8.6, VMM: yes Apr 17 23:09:42 athena ollama.listener[2840147]: Device 1: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes Apr 17 23:09:42 athena ollama.listener[2840147]: llm_load_tensors: ggml ctx size = 1.10 MiB Apr 17 23:09:46 athena ollama.listener[2840147]: llm_load_tensors: offloading 39 repeating layers to GPU Apr 17 23:09:46 athena ollama.listener[2840147]: llm_load_tensors: offloaded 39/41 layers to GPU Apr 17 23:09:46 athena ollama.listener[2840147]: llm_load_tensors: CPU buffer size = 70752.49 MiB Apr 17 23:09:46 athena ollama.listener[2840147]: llm_load_tensors: CUDA0 buffer size = 45460.59 MiB Apr 17 23:09:46 athena ollama.listener[2840147]: llm_load_tensors: CUDA1 buffer size = 22730.30 MiB Apr 17 23:09:53 athena ollama.listener[2840147]: .................................................................................................... Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: n_ctx = 2048 Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: n_batch = 512 Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: n_ubatch = 512 Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: freq_base = 500000.0 Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: freq_scale = 1 Apr 17 23:09:53 athena ollama.listener[2840147]: llama_kv_cache_init: CUDA_Host KV buffer size = 8.00 MiB Apr 17 23:09:53 athena ollama.listener[2840147]: llama_kv_cache_init: CUDA0 KV buffer size = 208.00 MiB Apr 17 23:09:53 athena ollama.listener[2840147]: llama_kv_cache_init: CUDA1 KV buffer size = 104.00 MiB Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: KV self size = 320.00 MiB, K (f16): 160.00 MiB, V (f16): 160.00 MiB Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: CUDA_Host output buffer size = 0.41 MiB Apr 17 23:09:53 athena ollama.listener[2840147]: ggml_backend_cuda_buffer_type_alloc_buffer: allocating 1794.00 MiB on device 0: cudaMalloc failed: out of memory Apr 17 23:09:53 athena ollama.listener[2840147]: ggml_gallocr_reserve_n: failed to allocate CUDA0 buffer of size 1881147392 Apr 17 23:09:53 athena ollama.listener[2840147]: llama_new_context_with_model: failed to allocate compute buffers Apr 17 23:09:53 athena ollama.listener[2840147]: llama_init_from_gpt_params: error: failed to create context with model '/media/data/ollama/blobs/sha256-1d12441f19436dbb0bcc4067e9d47921b944ef4a87b35873aa430e85e91a93c8' Apr 17 23:09:54 athena ollama.listener[2841851]: {"function":"load_model","level":"ERR","line":410,"model":"/media/data/ollama/blobs/sha256-1d12441f19436dbb0bcc4067e9d47921b944ef4a87b35873aa430e85e91a93c8","msg":"unable to load model","tid":"123385345642496","timestamp":1713384594} ``` When watching nvidia-smi (`watch nvidia-smi`) I see that the GPU 0 (the A6000) gets its memory nearly fully allocated, before the malloc failure begins. ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.32 (or current head of main branch)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3711/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3711/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3779
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3779/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3779/comments
https://api.github.com/repos/ollama/ollama/issues/3779/events
https://github.com/ollama/ollama/issues/3779
2,254,549,576
I_kwDOJ0Z1Ps6GYbJI
3,779
How to check memory utilization rate
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/taozhiyuai/followers", "following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}", "gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}", "starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions", "organizations_url": "https://api.github.com/users/taozhiyuai/orgs", "repos_url": "https://api.github.com/users/taozhiyuai/repos", "events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}", "received_events_url": "https://api.github.com/users/taozhiyuai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-04-20T13:19:09
2024-05-02T10:55:11
2024-05-02T10:55:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
after loading a model,How to check v-memory utilization rate
{ "login": "taozhiyuai", "id": 146583103, "node_id": "U_kgDOCLyuPw", "avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/taozhiyuai", "html_url": "https://github.com/taozhiyuai", "followers_url": "https://api.github.com/users/taozhiyuai/followers", "following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}", "gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}", "starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions", "organizations_url": "https://api.github.com/users/taozhiyuai/orgs", "repos_url": "https://api.github.com/users/taozhiyuai/repos", "events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}", "received_events_url": "https://api.github.com/users/taozhiyuai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3779/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3779/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/431
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/431/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/431/comments
https://api.github.com/repos/ollama/ollama/issues/431/events
https://github.com/ollama/ollama/issues/431
1,868,486,927
I_kwDOJ0Z1Ps5vXtkP
431
UTF-8 characters
{ "login": "arpecop", "id": 168974, "node_id": "MDQ6VXNlcjE2ODk3NA==", "avatar_url": "https://avatars.githubusercontent.com/u/168974?v=4", "gravatar_id": "", "url": "https://api.github.com/users/arpecop", "html_url": "https://github.com/arpecop", "followers_url": "https://api.github.com/users/arpecop/followers", "following_url": "https://api.github.com/users/arpecop/following{/other_user}", "gists_url": "https://api.github.com/users/arpecop/gists{/gist_id}", "starred_url": "https://api.github.com/users/arpecop/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/arpecop/subscriptions", "organizations_url": "https://api.github.com/users/arpecop/orgs", "repos_url": "https://api.github.com/users/arpecop/repos", "events_url": "https://api.github.com/users/arpecop/events{/privacy}", "received_events_url": "https://api.github.com/users/arpecop/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
4
2023-08-27T12:56:34
2023-08-30T21:11:50
2023-08-30T21:11:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How to render properly such characters , is this per-model issue ... or font related ? ![Screenshot 2023-08-27 at 15 54 34](https://github.com/jmorganca/ollama/assets/168974/6cc84d4a-f438-467f-bf7c-207ec45ee6fe)
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/431/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/431/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1546
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1546/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1546/comments
https://api.github.com/repos/ollama/ollama/issues/1546/events
https://github.com/ollama/ollama/issues/1546
2,044,086,055
I_kwDOJ0Z1Ps551kcn
1,546
Is running behind a proxy fully supported?
{ "login": "pardeike", "id": 853584, "node_id": "MDQ6VXNlcjg1MzU4NA==", "avatar_url": "https://avatars.githubusercontent.com/u/853584?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pardeike", "html_url": "https://github.com/pardeike", "followers_url": "https://api.github.com/users/pardeike/followers", "following_url": "https://api.github.com/users/pardeike/following{/other_user}", "gists_url": "https://api.github.com/users/pardeike/gists{/gist_id}", "starred_url": "https://api.github.com/users/pardeike/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pardeike/subscriptions", "organizations_url": "https://api.github.com/users/pardeike/orgs", "repos_url": "https://api.github.com/users/pardeike/repos", "events_url": "https://api.github.com/users/pardeike/events{/privacy}", "received_events_url": "https://api.github.com/users/pardeike/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2023-12-15T17:16:41
2023-12-28T12:17:29
2023-12-15T17:24:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, when I run with a configured proxy, I get the error `lookup registry.ollama.ai: no such host`: ``` > HTTPS_PROXY=http://REDACTED.XXX:10000 ollama run llama2 pulling manifest Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": dial tcp: lookup registry.ollama.ai: no such host ``` I then added a manual dns entry to /etc/hosts and ran the command again, but this time after trying to fetch the manifest, I get ``` > HTTPS_PROXY=http://REDACTED.XXX:10000 ollama run llama2 pulling manifest Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/llama2/manifests/latest": dial tcp 34.120.132.20:443: connect: network is unreachable ``` Going to those urls in Safari works just fine so it seems that the proxy support isn't fully working.
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1546/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1546/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1151
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1151/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1151/comments
https://api.github.com/repos/ollama/ollama/issues/1151/events
https://github.com/ollama/ollama/pull/1151
1,996,326,933
PR_kwDOJ0Z1Ps5fmfF0
1,151
install: fix enable contrib on debian 12
{ "login": "yanndegat", "id": 17246629, "node_id": "MDQ6VXNlcjE3MjQ2NjI5", "avatar_url": "https://avatars.githubusercontent.com/u/17246629?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yanndegat", "html_url": "https://github.com/yanndegat", "followers_url": "https://api.github.com/users/yanndegat/followers", "following_url": "https://api.github.com/users/yanndegat/following{/other_user}", "gists_url": "https://api.github.com/users/yanndegat/gists{/gist_id}", "starred_url": "https://api.github.com/users/yanndegat/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yanndegat/subscriptions", "organizations_url": "https://api.github.com/users/yanndegat/orgs", "repos_url": "https://api.github.com/users/yanndegat/repos", "events_url": "https://api.github.com/users/yanndegat/events{/privacy}", "received_events_url": "https://api.github.com/users/yanndegat/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-11-16T08:23:44
2023-11-16T20:53:07
2023-11-16T20:53:07
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1151", "html_url": "https://github.com/ollama/ollama/pull/1151", "diff_url": "https://github.com/ollama/ollama/pull/1151.diff", "patch_url": "https://github.com/ollama/ollama/pull/1151.patch", "merged_at": "2023-11-16T20:53:07" }
On debian 12, sources definitions have moved from /etc/apt/sources.list to /etc/apt/sources.list.d/debian.sources
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1151/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1151/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7571
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7571/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7571/comments
https://api.github.com/repos/ollama/ollama/issues/7571/events
https://github.com/ollama/ollama/issues/7571
2,643,262,341
I_kwDOJ0Z1Ps6djPuF
7,571
Please add Qwen-VL Qwen2.5-VL Qwen2-VL multimodal large model
{ "login": "smileyboy2019", "id": 59221294, "node_id": "MDQ6VXNlcjU5MjIxMjk0", "avatar_url": "https://avatars.githubusercontent.com/u/59221294?v=4", "gravatar_id": "", "url": "https://api.github.com/users/smileyboy2019", "html_url": "https://github.com/smileyboy2019", "followers_url": "https://api.github.com/users/smileyboy2019/followers", "following_url": "https://api.github.com/users/smileyboy2019/following{/other_user}", "gists_url": "https://api.github.com/users/smileyboy2019/gists{/gist_id}", "starred_url": "https://api.github.com/users/smileyboy2019/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/smileyboy2019/subscriptions", "organizations_url": "https://api.github.com/users/smileyboy2019/orgs", "repos_url": "https://api.github.com/users/smileyboy2019/repos", "events_url": "https://api.github.com/users/smileyboy2019/events{/privacy}", "received_events_url": "https://api.github.com/users/smileyboy2019/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-11-08T08:19:59
2024-11-12T00:40:52
2024-11-12T00:40:52
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7571/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7571/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4353
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4353/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4353/comments
https://api.github.com/repos/ollama/ollama/issues/4353/events
https://github.com/ollama/ollama/issues/4353
2,290,802,195
I_kwDOJ0Z1Ps6Iit4T
4,353
Unable to determine if Ollama can be run on GPU after running Linux commands
{ "login": "jerry123cao", "id": 124961434, "node_id": "U_kgDOB3LCmg", "avatar_url": "https://avatars.githubusercontent.com/u/124961434?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jerry123cao", "html_url": "https://github.com/jerry123cao", "followers_url": "https://api.github.com/users/jerry123cao/followers", "following_url": "https://api.github.com/users/jerry123cao/following{/other_user}", "gists_url": "https://api.github.com/users/jerry123cao/gists{/gist_id}", "starred_url": "https://api.github.com/users/jerry123cao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jerry123cao/subscriptions", "organizations_url": "https://api.github.com/users/jerry123cao/orgs", "repos_url": "https://api.github.com/users/jerry123cao/repos", "events_url": "https://api.github.com/users/jerry123cao/events{/privacy}", "received_events_url": "https://api.github.com/users/jerry123cao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-05-11T09:18:56
2024-05-21T23:21:51
2024-05-21T23:21:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello, I have a question. If I use the command curl - fsSL https://ollama.com/install.sh |Can Ollama run on GPU after installing the Nvidia driver and cuda toolkit in SH?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4353/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4353/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5292
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5292/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5292/comments
https://api.github.com/repos/ollama/ollama/issues/5292/events
https://github.com/ollama/ollama/issues/5292
2,374,357,161
I_kwDOJ0Z1Ps6NhdCp
5,292
请上架InternVL
{ "login": "enryteam", "id": 20081090, "node_id": "MDQ6VXNlcjIwMDgxMDkw", "avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4", "gravatar_id": "", "url": "https://api.github.com/users/enryteam", "html_url": "https://github.com/enryteam", "followers_url": "https://api.github.com/users/enryteam/followers", "following_url": "https://api.github.com/users/enryteam/following{/other_user}", "gists_url": "https://api.github.com/users/enryteam/gists{/gist_id}", "starred_url": "https://api.github.com/users/enryteam/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/enryteam/subscriptions", "organizations_url": "https://api.github.com/users/enryteam/orgs", "repos_url": "https://api.github.com/users/enryteam/repos", "events_url": "https://api.github.com/users/enryteam/events{/privacy}", "received_events_url": "https://api.github.com/users/enryteam/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2024-06-26T05:58:42
2025-01-28T13:33:00
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ERROR: type should be string, got "\r\nhttps://github.com/OpenGVLab/InternVL\r\n\r\nthanks 谢谢\r\n\r\n\r\n\r\nollama0.1.43\r\nerror format not yet support!\r\n错误格式尚不支持!\r\n错误格式尚不支持!\r\n搞了多次 均报错"
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5292/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5292/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/367
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/367/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/367/comments
https://api.github.com/repos/ollama/ollama/issues/367/events
https://github.com/ollama/ollama/pull/367
1,854,193,882
PR_kwDOJ0Z1Ps5YHWrd
367
fix nous-hermes model file size listing in readme
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-08-17T03:41:42
2023-08-17T03:42:01
2023-08-17T03:42:00
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/367", "html_url": "https://github.com/ollama/ollama/pull/367", "diff_url": "https://github.com/ollama/ollama/pull/367.diff", "patch_url": "https://github.com/ollama/ollama/pull/367.patch", "merged_at": "2023-08-17T03:42:00" }
fix nous-hermes model file size listing in readme
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/367/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/367/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6661
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6661/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6661/comments
https://api.github.com/repos/ollama/ollama/issues/6661/events
https://github.com/ollama/ollama/pull/6661
2,508,691,339
PR_kwDOJ0Z1Ps56kwXL
6,661
commit
{ "login": "rpreslar4765", "id": 89657947, "node_id": "MDQ6VXNlcjg5NjU3OTQ3", "avatar_url": "https://avatars.githubusercontent.com/u/89657947?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rpreslar4765", "html_url": "https://github.com/rpreslar4765", "followers_url": "https://api.github.com/users/rpreslar4765/followers", "following_url": "https://api.github.com/users/rpreslar4765/following{/other_user}", "gists_url": "https://api.github.com/users/rpreslar4765/gists{/gist_id}", "starred_url": "https://api.github.com/users/rpreslar4765/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rpreslar4765/subscriptions", "organizations_url": "https://api.github.com/users/rpreslar4765/orgs", "repos_url": "https://api.github.com/users/rpreslar4765/repos", "events_url": "https://api.github.com/users/rpreslar4765/events{/privacy}", "received_events_url": "https://api.github.com/users/rpreslar4765/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-09-05T20:53:42
2024-09-05T21:16:01
2024-09-05T21:16:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6661", "html_url": "https://github.com/ollama/ollama/pull/6661", "diff_url": "https://github.com/ollama/ollama/pull/6661.diff", "patch_url": "https://github.com/ollama/ollama/pull/6661.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6661/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6661/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8345
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8345/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8345/comments
https://api.github.com/repos/ollama/ollama/issues/8345/events
https://github.com/ollama/ollama/issues/8345
2,775,164,105
I_kwDOJ0Z1Ps6laaTJ
8,345
llama3.2-vision does not support tools
{ "login": "kwaa", "id": 50108258, "node_id": "MDQ6VXNlcjUwMTA4MjU4", "avatar_url": "https://avatars.githubusercontent.com/u/50108258?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kwaa", "html_url": "https://github.com/kwaa", "followers_url": "https://api.github.com/users/kwaa/followers", "following_url": "https://api.github.com/users/kwaa/following{/other_user}", "gists_url": "https://api.github.com/users/kwaa/gists{/gist_id}", "starred_url": "https://api.github.com/users/kwaa/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kwaa/subscriptions", "organizations_url": "https://api.github.com/users/kwaa/orgs", "repos_url": "https://api.github.com/users/kwaa/repos", "events_url": "https://api.github.com/users/kwaa/events{/privacy}", "received_events_url": "https://api.github.com/users/kwaa/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
17
2025-01-08T11:34:22
2025-01-18T09:30:37
2025-01-15T21:25:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I get above error when I use `llama3.2-vision:11b` with tool calling. I think `llama3.2-vision:11b` should support tool calls: https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/vision_prompt_format.md#builtin-and-zero-shot-tool-calling - [llama3.2:3b template](https://ollama.com/library/llama3.2/blobs/966de95ca8a6) - [llama3.2-vision:11b template](https://ollama.com/library/llama3.2-vision/blobs/715415638c9c) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.0
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8345/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8345/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1421
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1421/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1421/comments
https://api.github.com/repos/ollama/ollama/issues/1421/events
https://github.com/ollama/ollama/pull/1421
2,031,577,402
PR_kwDOJ0Z1Ps5hd4Rw
1,421
fix redundant newline
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-07T21:44:54
2023-12-07T21:47:24
2023-12-07T21:47:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1421", "html_url": "https://github.com/ollama/ollama/pull/1421", "diff_url": "https://github.com/ollama/ollama/pull/1421.diff", "patch_url": "https://github.com/ollama/ollama/pull/1421.patch", "merged_at": "2023-12-07T21:47:23" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1421/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1421/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8020
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8020/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8020/comments
https://api.github.com/repos/ollama/ollama/issues/8020/events
https://github.com/ollama/ollama/issues/8020
2,728,619,855
I_kwDOJ0Z1Ps6io29P
8,020
Nancy finds security vulnerabilities
{ "login": "mitar", "id": 585279, "node_id": "MDQ6VXNlcjU4NTI3OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/585279?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mitar", "html_url": "https://github.com/mitar", "followers_url": "https://api.github.com/users/mitar/followers", "following_url": "https://api.github.com/users/mitar/following{/other_user}", "gists_url": "https://api.github.com/users/mitar/gists{/gist_id}", "starred_url": "https://api.github.com/users/mitar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mitar/subscriptions", "organizations_url": "https://api.github.com/users/mitar/orgs", "repos_url": "https://api.github.com/users/mitar/repos", "events_url": "https://api.github.com/users/mitar/events{/privacy}", "received_events_url": "https://api.github.com/users/mitar/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
0
2024-12-10T00:29:13
2024-12-10T00:29:13
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? It finds: * [CVE-2024-8063](https://ossindex.sonatype.org/vulnerability/CVE-2024-8063?component-type=golang&component-name=github.com%2Follama%2Follama&utm_source=nancy-client&utm_medium=integration&utm_content=1.0.46) CWE-369: Divide By Zero * [CVE-2024-39719](https://ossindex.sonatype.org/vulnerability/CVE-2024-39719?component-type=golang&component-name=github.com%2Follama%2Follama&utm_source=nancy-client&utm_medium=integration&utm_content=1.0.46) CWE-209: Information Exposure Through an Error Message The latest stable Ollama version does not fix them, so I am opening an issue to track that. ### OS Linux ### GPU Intel ### CPU Intel ### Ollama version 0.5.1
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8020/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8020/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2467
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2467/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2467/comments
https://api.github.com/repos/ollama/ollama/issues/2467/events
https://github.com/ollama/ollama/pull/2467
2,130,595,963
PR_kwDOJ0Z1Ps5mqQ3_
2,467
Check image filetype in api handlers
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-12T17:17:19
2024-02-12T19:16:21
2024-02-12T19:16:20
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2467", "html_url": "https://github.com/ollama/ollama/pull/2467", "diff_url": "https://github.com/ollama/ollama/pull/2467.diff", "patch_url": "https://github.com/ollama/ollama/pull/2467.patch", "merged_at": "2024-02-12T19:16:20" }
Fixes: https://github.com/ollama/ollama/issues/2456
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2467/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2467/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1067
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1067/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1067/comments
https://api.github.com/repos/ollama/ollama/issues/1067/events
https://github.com/ollama/ollama/pull/1067
1,986,694,006
PR_kwDOJ0Z1Ps5fF3bO
1,067
add custom ollama-runner
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-11-10T01:45:17
2023-12-06T23:54:38
2023-11-21T20:14:50
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1067", "html_url": "https://github.com/ollama/ollama/pull/1067", "diff_url": "https://github.com/ollama/ollama/pull/1067.diff", "patch_url": "https://github.com/ollama/ollama/pull/1067.patch", "merged_at": null }
- update llama.cpp examples with custom ollama-runner - update llama.cpp gguf version to latest This change adds a custom inference server to llama.cpp based on the server we use in the current version, but with excess features removed. This allows us to have a more stable interface to build on when llama.cpp updates. To review this please pull down the changes run `go generate ./...` and review the contents of the `llm/llama.cpp/gguf/examples/ollama-runner` This change may be superseded by packaging in llama.cpp directly in the near future.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1067/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1420
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1420/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1420/comments
https://api.github.com/repos/ollama/ollama/issues/1420/events
https://github.com/ollama/ollama/pull/1420
2,031,437,051
PR_kwDOJ0Z1Ps5hdZct
1,420
os specific ctrl-z
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-12-07T20:03:08
2023-12-14T06:46:26
2023-12-11T15:48:15
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1420", "html_url": "https://github.com/ollama/ollama/pull/1420", "diff_url": "https://github.com/ollama/ollama/pull/1420.diff", "patch_url": "https://github.com/ollama/ollama/pull/1420.patch", "merged_at": "2023-12-11T15:48:15" }
Add OS specific readline functions. Windows does not support these suspend system calls, so make ctrl-z a no-op on windows. This fixes development windows native builds. resolves #1414
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1420/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1420/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7824
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7824/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7824/comments
https://api.github.com/repos/ollama/ollama/issues/7824/events
https://github.com/ollama/ollama/issues/7824
2,688,581,053
I_kwDOJ0Z1Ps6gQH29
7,824
Tool calling parsing for llama3.2
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2024-11-24T23:12:47
2024-11-24T23:12:47
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Llama 3.2 tool call outputs [are not in JSON](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/text_prompt_format.md) and so Ollama's tool parsing needs to be updated ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7824/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7824/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2586
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2586/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2586/comments
https://api.github.com/repos/ollama/ollama/issues/2586/events
https://github.com/ollama/ollama/issues/2586
2,141,379,273
I_kwDOJ0Z1Ps5_otrJ
2,586
Question on model updating
{ "login": "owenzhao", "id": 2182896, "node_id": "MDQ6VXNlcjIxODI4OTY=", "avatar_url": "https://avatars.githubusercontent.com/u/2182896?v=4", "gravatar_id": "", "url": "https://api.github.com/users/owenzhao", "html_url": "https://github.com/owenzhao", "followers_url": "https://api.github.com/users/owenzhao/followers", "following_url": "https://api.github.com/users/owenzhao/following{/other_user}", "gists_url": "https://api.github.com/users/owenzhao/gists{/gist_id}", "starred_url": "https://api.github.com/users/owenzhao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/owenzhao/subscriptions", "organizations_url": "https://api.github.com/users/owenzhao/orgs", "repos_url": "https://api.github.com/users/owenzhao/repos", "events_url": "https://api.github.com/users/owenzhao/events{/privacy}", "received_events_url": "https://api.github.com/users/owenzhao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
2
2024-02-19T02:42:14
2024-03-11T20:45:20
2024-03-11T20:45:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I know I could download the latest model when a model is not installed. Say: ```swift ollama run llama2 ``` After some days, say one month, the model will be updated. Can ollama updates to the latest model automatically? I didn't find a way to do that. Please help.
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2586/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2586/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2247
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2247/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2247/comments
https://api.github.com/repos/ollama/ollama/issues/2247/events
https://github.com/ollama/ollama/pull/2247
2,104,479,078
PR_kwDOJ0Z1Ps5lRYk5
2,247
Update import instructions to use convert and quantize tooling from llama.cpp submodule
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-29T01:45:12
2024-02-05T05:50:45
2024-02-05T05:50:44
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2247", "html_url": "https://github.com/ollama/ollama/pull/2247", "diff_url": "https://github.com/ollama/ollama/pull/2247.diff", "patch_url": "https://github.com/ollama/ollama/pull/2247.patch", "merged_at": "2024-02-05T05:50:44" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2247/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2247/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5994
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5994/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5994/comments
https://api.github.com/repos/ollama/ollama/issues/5994/events
https://github.com/ollama/ollama/pull/5994
2,432,953,146
PR_kwDOJ0Z1Ps52nmhq
5,994
server: fix race conditions during download
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-26T20:44:18
2024-07-26T21:24:25
2024-07-26T21:24:24
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5994", "html_url": "https://github.com/ollama/ollama/pull/5994", "diff_url": "https://github.com/ollama/ollama/pull/5994.diff", "patch_url": "https://github.com/ollama/ollama/pull/5994.patch", "merged_at": "2024-07-26T21:24:24" }
This fixes various data races scattered throughout the download/pull client where the client was accessing the download state concurrently. This commit is mostly a hot-fix and will be replaced by a new client one day soon. Also, remove the unnecessary opts argument from downloadChunk.
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5994/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5994/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3839
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3839/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3839/comments
https://api.github.com/repos/ollama/ollama/issues/3839/events
https://github.com/ollama/ollama/issues/3839
2,258,206,018
I_kwDOJ0Z1Ps6GmX1C
3,839
Feature Request: Detect Truncation Due to Exceeding Context Size
{ "login": "guoxf", "id": 3947566, "node_id": "MDQ6VXNlcjM5NDc1NjY=", "avatar_url": "https://avatars.githubusercontent.com/u/3947566?v=4", "gravatar_id": "", "url": "https://api.github.com/users/guoxf", "html_url": "https://github.com/guoxf", "followers_url": "https://api.github.com/users/guoxf/followers", "following_url": "https://api.github.com/users/guoxf/following{/other_user}", "gists_url": "https://api.github.com/users/guoxf/gists{/gist_id}", "starred_url": "https://api.github.com/users/guoxf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/guoxf/subscriptions", "organizations_url": "https://api.github.com/users/guoxf/orgs", "repos_url": "https://api.github.com/users/guoxf/repos", "events_url": "https://api.github.com/users/guoxf/events{/privacy}", "received_events_url": "https://api.github.com/users/guoxf/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
2
2024-04-23T07:54:37
2024-07-01T17:15:46
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Understanding whether model output has been truncated due to exceeding context size is crucial for trusting the model to provide complete and accurate information. Here are some specific examples that illustrate why it's necessary to know if the output has been truncated: 1. **Question-Answering Systems**: When building a question-answering system, if the answer to a user's query is truncated due to exceeding context size, the user may receive incomplete or inaccurate answers, which can degrade user experience and system reliability. 2. **Text Summarization**: When using a model to generate summaries of articles or reports, if the summary is truncated due to exceeding context size, it may omit key information, leading to incomplete understanding of the original content. 3. **Chatbots**: In chatbot applications, if a response in a conversation is truncated due to exceeding context size, it may cause the dialogue flow to be interrupted, affecting the user's interaction experience with the bot. 4. **Content Generation**: When using a model to generate articles, stories, or other creative content, if the generated text is truncated due to exceeding context size, it may disrupt the coherence and completeness of the content. 5. **Machine Translation**: In machine translation scenarios, if the translated text is truncated due to exceeding context size, it may result in the loss of the latter part of the translation, affecting translation accuracy. 6. **Natural Language Processing Tasks**: When dealing with long texts in any natural language processing task, such as sentiment analysis or topic classification, if the model's output is truncated due to exceeding context size, it may lead to incorrect processing results. 7. **Legal and Compliance**: In applications with high legal or compliance requirements, the completeness and accuracy of information are crucial. If model output is truncated due to exceeding context size, it may violate these requirements and lead to serious consequences. 8. **Academic and Research**: When using large language models in academic research, ensuring the completeness and accuracy of results is vital for validating research hypotheses and conclusions. If model output is truncated, it may negatively impact the effectiveness and reliability of the research. Please consider implementing a mechanism within the API that clearly indicates whether the output has been truncated. This could be a boolean flag in the response payload or an error message that provides insight into the truncation. Thank you for your attention to this matter. I believe this feature would greatly enhance the usability and trustworthiness of ollama for all users.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3839/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3839/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6044
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6044/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6044/comments
https://api.github.com/repos/ollama/ollama/issues/6044/events
https://github.com/ollama/ollama/issues/6044
2,435,056,414
I_kwDOJ0Z1Ps6RJAMe
6,044
Illegal istruction in ollama_llama_server runner
{ "login": "SnowyCoder", "id": 10945206, "node_id": "MDQ6VXNlcjEwOTQ1MjA2", "avatar_url": "https://avatars.githubusercontent.com/u/10945206?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SnowyCoder", "html_url": "https://github.com/SnowyCoder", "followers_url": "https://api.github.com/users/SnowyCoder/followers", "following_url": "https://api.github.com/users/SnowyCoder/following{/other_user}", "gists_url": "https://api.github.com/users/SnowyCoder/gists{/gist_id}", "starred_url": "https://api.github.com/users/SnowyCoder/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SnowyCoder/subscriptions", "organizations_url": "https://api.github.com/users/SnowyCoder/orgs", "repos_url": "https://api.github.com/users/SnowyCoder/repos", "events_url": "https://api.github.com/users/SnowyCoder/events{/privacy}", "received_events_url": "https://api.github.com/users/SnowyCoder/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-07-29T10:34:52
2024-07-29T10:46:24
2024-07-29T10:46:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I tried to run llama3 model with ollama. Reproduction (with my CPU: `AMD Ryzen 7 7735HS`) 1. Start server: `ollama serve` 2. Try to run llama3: `ollama run llama3 (The same error occurs with llama3.1) The server will execute a new runner with the following arguments: ```/tmp/ollama996131774/runners/cpu/ollama_llama_server --model /home/snowy/.ollama/models/blobs/ sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa --ctx-size 8192 --batch-size 512 --embedding --log-disable --no-mmap --parallel 4 --port 42857``` And the model will fail with the following errors: ``` llama runner process has terminated: signal: illegal instruction (core dumped) ``` When opening the runner with GDB, the program crashes when trying to run the following instruction: `vmovupd %zmm0,0x13(%rax)` (ggml_init+775) that requires AVX512F feature flag. Complete log: ``` ❯ ollama serve 2024/07/29 12:09:49 routes.go:1099: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/home/snowy/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_RUNNERS_DIR: OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR: ROCR_VISIBLE_DEVICES:]" time=2024-07-29T12:09:49.854+02:00 level=INFO source=images.go:784 msg="total blobs: 10" time=2024-07-29T12:09:49.855+02:00 level=INFO source=images.go:791 msg="total unused blobs removed: 0" time=2024-07-29T12:09:49.855+02:00 level=INFO source=routes.go:1146 msg="Listening on 127.0.0.1:11434 (version 0.3.0)" time=2024-07-29T12:09:49.855+02:00 level=INFO source=payload.go:30 msg="extracting embedded files" dir=/tmp/ollama996131774/runners time=2024-07-29T12:09:57.575+02:00 level=INFO source=payload.go:44 msg="Dynamic LLM libraries [cpu rocm]" time=2024-07-29T12:09:57.575+02:00 level=INFO source=gpu.go:205 msg="looking for compatible GPUs" time=2024-07-29T12:09:57.657+02:00 level=WARN source=amd_linux.go:58 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory" time=2024-07-29T12:09:57.662+02:00 level=WARN source=amd_linux.go:325 msg="amdgpu is not supported" gpu=0 gpu_type=gfx1035 library=/opt/rocm/lib supported_types="[gfx1030 gfx1100 gfx1101 gfx1102 gfx900 gfx906 gfx908 gfx90a gfx940 gfx941 gfx942]" time=2024-07-29T12:09:57.662+02:00 level=WARN source=amd_linux.go:327 msg="See https://github.com/ollama/ollama/blob/main/docs/gpu.md#overrides for HSA_OVERRIDE_GFX_VERSION usage" time=2024-07-29T12:09:57.662+02:00 level=INFO source=amd_linux.go:345 msg="no compatible amdgpu devices detected" time=2024-07-29T12:09:57.662+02:00 level=INFO source=gpu.go:346 msg="no compatible GPUs were discovered" time=2024-07-29T12:09:57.662+02:00 level=INFO source=types.go:105 msg="inference compute" id=0 library=cpu compute="" driver=0.0 name="" total="13.3 GiB" available="9.1 GiB" [GIN] 2024/07/29 - 12:09:57 | 200 | 66.764µs | 127.0.0.1 | HEAD "/" [GIN] 2024/07/29 - 12:09:57 | 200 | 41.795654ms | 127.0.0.1 | POST "/api/show" time=2024-07-29T12:09:57.790+02:00 level=INFO source=memory.go:309 msg="offload to cpu" layers.requested=-1 layers.model=33 layers.offload=0 layers.split="" memory.available="[9.1 GiB]" memory.required.full="5.8 GiB" memory.required.partial="0 B" memory.required.kv="1.0 GiB" memory.required.allocations="[5.8 GiB]" memory.weights.total="4.7 GiB" memory.weights.repeating="4.3 GiB" memory.weights.nonrepeating="411.0 MiB" memory.graph.full="560.0 MiB" memory.graph.partial="677.5 MiB" time=2024-07-29T12:09:57.792+02:00 level=INFO source=server.go:383 msg="starting llama server" cmd="/tmp/ollama996131774/runners/cpu/ollama_llama_server --model /home/snowy/.ollama/models/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa --ctx-size 8192 --batch-size 512 --embedding --log-disable --no-mmap --parallel 4 --port 42857" time=2024-07-29T12:09:57.793+02:00 level=INFO source=sched.go:437 msg="loaded runners" count=1 time=2024-07-29T12:09:57.793+02:00 level=INFO source=server.go:583 msg="waiting for llama runner to start responding" time=2024-07-29T12:09:57.794+02:00 level=INFO source=server.go:617 msg="waiting for server to become available" status="llm server error" time=2024-07-29T12:09:58.044+02:00 level=ERROR source=sched.go:443 msg="error loading llama server" error="llama runner process has terminated: signal: illegal instruction (core dumped)" [GIN] 2024/07/29 - 12:09:58 | 500 | 338.214098ms | 127.0.0.1 | POST "/api/chat" [GIN] 2024/07/29 - 12:18:18 | 200 | 63.63µs | 127.0.0.1 | GET "/api/version" ``` [coredump.zip](https://github.com/user-attachments/files/16411350/coredump.zip) [cpuinfo.txt](https://github.com/user-attachments/files/16411356/cpuinfo.txt) ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.3.0
{ "login": "SnowyCoder", "id": 10945206, "node_id": "MDQ6VXNlcjEwOTQ1MjA2", "avatar_url": "https://avatars.githubusercontent.com/u/10945206?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SnowyCoder", "html_url": "https://github.com/SnowyCoder", "followers_url": "https://api.github.com/users/SnowyCoder/followers", "following_url": "https://api.github.com/users/SnowyCoder/following{/other_user}", "gists_url": "https://api.github.com/users/SnowyCoder/gists{/gist_id}", "starred_url": "https://api.github.com/users/SnowyCoder/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/SnowyCoder/subscriptions", "organizations_url": "https://api.github.com/users/SnowyCoder/orgs", "repos_url": "https://api.github.com/users/SnowyCoder/repos", "events_url": "https://api.github.com/users/SnowyCoder/events{/privacy}", "received_events_url": "https://api.github.com/users/SnowyCoder/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6044/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6044/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2368
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2368/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2368/comments
https://api.github.com/repos/ollama/ollama/issues/2368/events
https://github.com/ollama/ollama/pull/2368
2,119,989,649
PR_kwDOJ0Z1Ps5mGd7d
2,368
Replace `reflect` usage in option parsing
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-02-06T05:25:58
2024-08-11T16:52:11
2024-08-11T16:52:11
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2368", "html_url": "https://github.com/ollama/ollama/pull/2368", "diff_url": "https://github.com/ollama/ollama/pull/2368.diff", "patch_url": "https://github.com/ollama/ollama/pull/2368.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2368/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2368/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6929
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6929/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6929/comments
https://api.github.com/repos/ollama/ollama/issues/6929/events
https://github.com/ollama/ollama/issues/6929
2,544,478,139
I_kwDOJ0Z1Ps6Xqae7
6,929
support qwen2.5:72b-instruct function call
{ "login": "chenfolingithub", "id": 9414696, "node_id": "MDQ6VXNlcjk0MTQ2OTY=", "avatar_url": "https://avatars.githubusercontent.com/u/9414696?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chenfolingithub", "html_url": "https://github.com/chenfolingithub", "followers_url": "https://api.github.com/users/chenfolingithub/followers", "following_url": "https://api.github.com/users/chenfolingithub/following{/other_user}", "gists_url": "https://api.github.com/users/chenfolingithub/gists{/gist_id}", "starred_url": "https://api.github.com/users/chenfolingithub/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chenfolingithub/subscriptions", "organizations_url": "https://api.github.com/users/chenfolingithub/orgs", "repos_url": "https://api.github.com/users/chenfolingithub/repos", "events_url": "https://api.github.com/users/chenfolingithub/events{/privacy}", "received_events_url": "https://api.github.com/users/chenfolingithub/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-09-24T06:36:16
2024-11-06T00:18:09
2024-11-06T00:18:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
qwen2.5:72b-instruct now support function call Returned its own unique data format, but currently ollama 0.3.11 does not correctly return formatted data for function calls
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6929/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6929/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4588
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4588/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4588/comments
https://api.github.com/repos/ollama/ollama/issues/4588/events
https://github.com/ollama/ollama/issues/4588
2,312,622,323
I_kwDOJ0Z1Ps6J19Dz
4,588
Tool calling feature support
{ "login": "Goekdeniz-Guelmez", "id": 60228478, "node_id": "MDQ6VXNlcjYwMjI4NDc4", "avatar_url": "https://avatars.githubusercontent.com/u/60228478?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Goekdeniz-Guelmez", "html_url": "https://github.com/Goekdeniz-Guelmez", "followers_url": "https://api.github.com/users/Goekdeniz-Guelmez/followers", "following_url": "https://api.github.com/users/Goekdeniz-Guelmez/following{/other_user}", "gists_url": "https://api.github.com/users/Goekdeniz-Guelmez/gists{/gist_id}", "starred_url": "https://api.github.com/users/Goekdeniz-Guelmez/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Goekdeniz-Guelmez/subscriptions", "organizations_url": "https://api.github.com/users/Goekdeniz-Guelmez/orgs", "repos_url": "https://api.github.com/users/Goekdeniz-Guelmez/repos", "events_url": "https://api.github.com/users/Goekdeniz-Guelmez/events{/privacy}", "received_events_url": "https://api.github.com/users/Goekdeniz-Guelmez/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-05-23T11:06:39
2024-05-23T21:39:55
2024-05-23T21:39:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
since there are more and more models that support function/tool calling. I think it would be a great feature for that support. It can be hard because the models could need Special tokens. My thought is that you can be able to set these spetial tokens or customizations in the Modelfile, something like: ```text PARAMETER tool_calls_start "" PARAMETER tool_calls_end "" PARAMETER tool_results_start "" PARAMETER tool_results_end "" ... ``` or (my preferred one): ```text TOOL_CALLS """"[TOOL_CALLS] {response}""" AVAILABLE_TOOLS """[AVAILABLE_TOOLS] {available tools list} [/AVAILABLE_TOOLS]""" TOOL_RESULTS """[TOOL_RESULTS] {tool response} [/TOOL_RESULTS]"""" ``` The API could be the same with OpenAI's or something different.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4588/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4588/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2424
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2424/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2424/comments
https://api.github.com/repos/ollama/ollama/issues/2424/events
https://github.com/ollama/ollama/issues/2424
2,126,622,691
I_kwDOJ0Z1Ps5-wa_j
2,424
Always getting a timeout error while querying using mistral using Ollama
{ "login": "Chakit22", "id": 118890138, "node_id": "U_kgDOBxYemg", "avatar_url": "https://avatars.githubusercontent.com/u/118890138?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Chakit22", "html_url": "https://github.com/Chakit22", "followers_url": "https://api.github.com/users/Chakit22/followers", "following_url": "https://api.github.com/users/Chakit22/following{/other_user}", "gists_url": "https://api.github.com/users/Chakit22/gists{/gist_id}", "starred_url": "https://api.github.com/users/Chakit22/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Chakit22/subscriptions", "organizations_url": "https://api.github.com/users/Chakit22/orgs", "repos_url": "https://api.github.com/users/Chakit22/repos", "events_url": "https://api.github.com/users/Chakit22/events{/privacy}", "received_events_url": "https://api.github.com/users/Chakit22/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
19
2024-02-09T07:36:05
2025-01-16T09:16:45
2024-05-07T23:53:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
``` Traceback (most recent call last): File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_exceptions.py", line 10, in map_exceptions yield File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_backends/sync.py", line 126, in read return self._sock.recv(max_bytes) ^^^^^^^^^^^^^^^^^^^^^^^^^^ TimeoutError: timed out The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_transports/default.py", line 67, in map_httpcore_exceptions yield File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_transports/default.py", line 231, in handle_request resp = self._pool.handle_request(req) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/connection_pool.py", line 268, in handle_request raise exc File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/connection_pool.py", line 251, in handle_request response = connection.handle_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/connection.py", line 103, in handle_request return self._connection.handle_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/http11.py", line 133, in handle_request raise exc File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/http11.py", line 111, in handle_request ) = self._receive_response_headers(**kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/http11.py", line 176, in _receive_response_headers event = self._receive_event(timeout=timeout) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_sync/http11.py", line 212, in _receive_event data = self._network_stream.read( ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_backends/sync.py", line 124, in read with map_exceptions(exc_map): File "/opt/homebrew/Cellar/python@3.11/3.11.7_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpcore/_exceptions.py", line 14, in map_exceptions raise to_exc(exc) from exc httpcore.ReadTimeout: timed out The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/Users/chakitrocks/Desktop/llm/index.py", line 57, in <module> response = query_engine.query("What does the author think about Star Trek? Give details.") ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/core/base_query_engine.py", line 40, in query return self._query(str_or_query_bundle) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/query_engine/retriever_query_engine.py", line 172, in _query response = self._response_synthesizer.synthesize( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/response_synthesizers/base.py", line 168, in synthesize response_str = self.get_response( ^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/response_synthesizers/compact_and_refine.py", line 38, in get_response return super().get_response( ^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/response_synthesizers/refine.py", line 146, in get_response response = self._give_response_single( ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/response_synthesizers/refine.py", line 202, in _give_response_single program( File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/response_synthesizers/refine.py", line 64, in __call__ answer = self._llm.predict( ^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/llms/llm.py", line 239, in predict chat_response = self.chat(messages) ^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/llms/base.py", line 100, in wrapped_llm_chat f_return_val = f(_self, messages, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/llama_index/llms/ollama.py", line 102, in chat response = client.post( ^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_client.py", line 1146, in post return self.request( ^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_client.py", line 828, in request return self.send(request, auth=auth, follow_redirects=follow_redirects) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_client.py", line 915, in send response = self._send_handling_auth( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_client.py", line 943, in _send_handling_auth response = self._send_handling_redirects( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_client.py", line 980, in _send_handling_redirects response = self._send_single_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_client.py", line 1016, in _send_single_request response = transport.handle_request(request) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_transports/default.py", line 230, in handle_request with map_httpcore_exceptions(): File "/opt/homebrew/Cellar/python@3.11/3.11.7_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/Users/chakitrocks/Desktop/llm/env/lib/python3.11/site-packages/httpx/_transports/default.py", line 84, in map_httpcore_exceptions raise mapped_exc(message) from exc httpx.ReadTimeout: timed out ``` I seem to get this error while I am trying to fetch the top 20 values using VectorStoreIndex. Here's the link to the blog I was trying to implement: https://blog.llamaindex.ai/running-mixtral-8x7-locally-with-llamaindex-e6cebeabe0ab I am getting a timeout while I am querying with `similarity_top_k=20`. What is the workaround this?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2424/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2424/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1025
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1025/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1025/comments
https://api.github.com/repos/ollama/ollama/issues/1025/events
https://github.com/ollama/ollama/issues/1025
1,980,404,467
I_kwDOJ0Z1Ps52CpLz
1,025
How to stop a model aftering it's served?
{ "login": "paulwongx", "id": 20178251, "node_id": "MDQ6VXNlcjIwMTc4MjUx", "avatar_url": "https://avatars.githubusercontent.com/u/20178251?v=4", "gravatar_id": "", "url": "https://api.github.com/users/paulwongx", "html_url": "https://github.com/paulwongx", "followers_url": "https://api.github.com/users/paulwongx/followers", "following_url": "https://api.github.com/users/paulwongx/following{/other_user}", "gists_url": "https://api.github.com/users/paulwongx/gists{/gist_id}", "starred_url": "https://api.github.com/users/paulwongx/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/paulwongx/subscriptions", "organizations_url": "https://api.github.com/users/paulwongx/orgs", "repos_url": "https://api.github.com/users/paulwongx/repos", "events_url": "https://api.github.com/users/paulwongx/events{/privacy}", "received_events_url": "https://api.github.com/users/paulwongx/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-11-07T02:12:01
2023-11-07T14:11:09
2023-11-07T14:11:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When the modelling is already being served, what is the command to stop it so that I can serve it again with different variables? I want to run `OLLAMA_ORIGINS=https://XXXXX ollama serve` but I'm getting an error that the address is already in use. Running `ollama --help` doesn't show a command to unserve...? tldr; how can I unserve?
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1025/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1025/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4544
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4544/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4544/comments
https://api.github.com/repos/ollama/ollama/issues/4544/events
https://github.com/ollama/ollama/pull/4544
2,306,532,844
PR_kwDOJ0Z1Ps5v_RLM
4,544
update langchain-python-rag-document example to use latest langchain apis
{ "login": "pixelsoccupied", "id": 19938874, "node_id": "MDQ6VXNlcjE5OTM4ODc0", "avatar_url": "https://avatars.githubusercontent.com/u/19938874?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pixelsoccupied", "html_url": "https://github.com/pixelsoccupied", "followers_url": "https://api.github.com/users/pixelsoccupied/followers", "following_url": "https://api.github.com/users/pixelsoccupied/following{/other_user}", "gists_url": "https://api.github.com/users/pixelsoccupied/gists{/gist_id}", "starred_url": "https://api.github.com/users/pixelsoccupied/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pixelsoccupied/subscriptions", "organizations_url": "https://api.github.com/users/pixelsoccupied/orgs", "repos_url": "https://api.github.com/users/pixelsoccupied/repos", "events_url": "https://api.github.com/users/pixelsoccupied/events{/privacy}", "received_events_url": "https://api.github.com/users/pixelsoccupied/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2024-05-20T18:27:53
2024-10-29T07:15:48
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4544", "html_url": "https://github.com/ollama/ollama/pull/4544", "diff_url": "https://github.com/ollama/ollama/pull/4544.diff", "patch_url": "https://github.com/ollama/ollama/pull/4544.patch", "merged_at": null }
Hello! Langchain library went through a lot changes since this helpful example was first added and users will find that this will not work out of box anymore. I have updated this to work with all the latest changes and also removed any deprecated functions. Hope others find this useful!
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4544/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4544/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3174
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3174/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3174/comments
https://api.github.com/repos/ollama/ollama/issues/3174/events
https://github.com/ollama/ollama/pull/3174
2,189,661,685
PR_kwDOJ0Z1Ps5pzx1y
3,174
llama: remove server static assets
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-16T00:42:03
2024-03-16T02:24:13
2024-03-16T02:24:12
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3174", "html_url": "https://github.com/ollama/ollama/pull/3174", "diff_url": "https://github.com/ollama/ollama/pull/3174.diff", "patch_url": "https://github.com/ollama/ollama/pull/3174.patch", "merged_at": "2024-03-16T02:24:12" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3174/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3174/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1439
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1439/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1439/comments
https://api.github.com/repos/ollama/ollama/issues/1439/events
https://github.com/ollama/ollama/issues/1439
2,033,191,401
I_kwDOJ0Z1Ps55MAnp
1,439
Serious bug
{ "login": "m-c-frank", "id": 61345033, "node_id": "MDQ6VXNlcjYxMzQ1MDMz", "avatar_url": "https://avatars.githubusercontent.com/u/61345033?v=4", "gravatar_id": "", "url": "https://api.github.com/users/m-c-frank", "html_url": "https://github.com/m-c-frank", "followers_url": "https://api.github.com/users/m-c-frank/followers", "following_url": "https://api.github.com/users/m-c-frank/following{/other_user}", "gists_url": "https://api.github.com/users/m-c-frank/gists{/gist_id}", "starred_url": "https://api.github.com/users/m-c-frank/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/m-c-frank/subscriptions", "organizations_url": "https://api.github.com/users/m-c-frank/orgs", "repos_url": "https://api.github.com/users/m-c-frank/repos", "events_url": "https://api.github.com/users/m-c-frank/events{/privacy}", "received_events_url": "https://api.github.com/users/m-c-frank/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-08T19:20:22
2023-12-08T19:23:58
2023-12-08T19:23:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
(base) mcfrank@interface:~$ orca show The following are not valid: show Warning: Could not load keyboard geometry for :0 BadName (named color or font does not exist) Resulting keymap file will not describe geometry ^C(base) mcfrank@interface:~$
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1439/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1439/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7411
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7411/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7411/comments
https://api.github.com/repos/ollama/ollama/issues/7411/events
https://github.com/ollama/ollama/pull/7411
2,622,487,493
PR_kwDOJ0Z1Ps6ATg6p
7,411
add more tests for getting the optimal tiled canvas
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-29T23:02:00
2024-10-29T23:28:04
2024-10-29T23:28:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7411", "html_url": "https://github.com/ollama/ollama/pull/7411", "diff_url": "https://github.com/ollama/ollama/pull/7411.diff", "patch_url": "https://github.com/ollama/ollama/pull/7411.patch", "merged_at": "2024-10-29T23:28:02" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7411/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7411/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/291
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/291/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/291/comments
https://api.github.com/repos/ollama/ollama/issues/291/events
https://github.com/ollama/ollama/issues/291
1,837,538,185
I_kwDOJ0Z1Ps5thpuJ
291
add `ollama serve` to the quickstart doc
{ "login": "marafa-sugarcrm", "id": 69474017, "node_id": "MDQ6VXNlcjY5NDc0MDE3", "avatar_url": "https://avatars.githubusercontent.com/u/69474017?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marafa-sugarcrm", "html_url": "https://github.com/marafa-sugarcrm", "followers_url": "https://api.github.com/users/marafa-sugarcrm/followers", "following_url": "https://api.github.com/users/marafa-sugarcrm/following{/other_user}", "gists_url": "https://api.github.com/users/marafa-sugarcrm/gists{/gist_id}", "starred_url": "https://api.github.com/users/marafa-sugarcrm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marafa-sugarcrm/subscriptions", "organizations_url": "https://api.github.com/users/marafa-sugarcrm/orgs", "repos_url": "https://api.github.com/users/marafa-sugarcrm/repos", "events_url": "https://api.github.com/users/marafa-sugarcrm/events{/privacy}", "received_events_url": "https://api.github.com/users/marafa-sugarcrm/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-08-05T02:51:59
2023-08-05T02:53:17
2023-08-05T02:53:17
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
i downloaded using `brew` and attempted to follow the quick start doc it is missing `ollama serve`
{ "login": "marafa-sugarcrm", "id": 69474017, "node_id": "MDQ6VXNlcjY5NDc0MDE3", "avatar_url": "https://avatars.githubusercontent.com/u/69474017?v=4", "gravatar_id": "", "url": "https://api.github.com/users/marafa-sugarcrm", "html_url": "https://github.com/marafa-sugarcrm", "followers_url": "https://api.github.com/users/marafa-sugarcrm/followers", "following_url": "https://api.github.com/users/marafa-sugarcrm/following{/other_user}", "gists_url": "https://api.github.com/users/marafa-sugarcrm/gists{/gist_id}", "starred_url": "https://api.github.com/users/marafa-sugarcrm/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/marafa-sugarcrm/subscriptions", "organizations_url": "https://api.github.com/users/marafa-sugarcrm/orgs", "repos_url": "https://api.github.com/users/marafa-sugarcrm/repos", "events_url": "https://api.github.com/users/marafa-sugarcrm/events{/privacy}", "received_events_url": "https://api.github.com/users/marafa-sugarcrm/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/291/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/291/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5881
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5881/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5881/comments
https://api.github.com/repos/ollama/ollama/issues/5881/events
https://github.com/ollama/ollama/issues/5881
2,425,609,012
I_kwDOJ0Z1Ps6Qk9s0
5,881
Is llama 3.1 already supported (on 2.8) or should we wait another update ?
{ "login": "Qualzz", "id": 35169816, "node_id": "MDQ6VXNlcjM1MTY5ODE2", "avatar_url": "https://avatars.githubusercontent.com/u/35169816?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Qualzz", "html_url": "https://github.com/Qualzz", "followers_url": "https://api.github.com/users/Qualzz/followers", "following_url": "https://api.github.com/users/Qualzz/following{/other_user}", "gists_url": "https://api.github.com/users/Qualzz/gists{/gist_id}", "starred_url": "https://api.github.com/users/Qualzz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Qualzz/subscriptions", "organizations_url": "https://api.github.com/users/Qualzz/orgs", "repos_url": "https://api.github.com/users/Qualzz/repos", "events_url": "https://api.github.com/users/Qualzz/events{/privacy}", "received_events_url": "https://api.github.com/users/Qualzz/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
20
2024-07-23T16:22:57
2024-09-02T00:15:39
2024-09-02T00:15:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The model page seems to already exists in ollama website, but the model is clearly behaving erratically, which makes me wonder if we should wait for an update before using llama 3.1. ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 2.8
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5881/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5881/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1445
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1445/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1445/comments
https://api.github.com/repos/ollama/ollama/issues/1445/events
https://github.com/ollama/ollama/pull/1445
2,033,535,254
PR_kwDOJ0Z1Ps5hkiJ6
1,445
fix: parallel queueing race condition caused silent failure
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-12-09T01:58:05
2023-12-09T19:14:03
2023-12-09T19:14:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1445", "html_url": "https://github.com/ollama/ollama/pull/1445", "diff_url": "https://github.com/ollama/ollama/pull/1445.diff", "patch_url": "https://github.com/ollama/ollama/pull/1445.patch", "merged_at": "2023-12-09T19:14:02" }
As of the most recent llama.cpp update concurrent requests had a race condition that would result in an empty response. This was not easy to observe since the response from the llm runner subprocess was a 200 with the error {"content":"slot unavailable"} in the response stream, which just silently closed the channel. This change resolves this by allowing multiple slots in the llm runner subprocess. We manage the queueing ourselves so this should be ok. @dhiltgen this may be a case we need to account for in the cgo changes.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1445/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1445/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3483
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3483/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3483/comments
https://api.github.com/repos/ollama/ollama/issues/3483/events
https://github.com/ollama/ollama/issues/3483
2,224,525,531
I_kwDOJ0Z1Ps6El5Db
3,483
Ollama hangs on CUDA devices when running multi-modal models
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-04-04T05:24:22
2024-04-04T06:30:58
2024-04-04T06:30:58
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ``` Apr 04 05:15:04 gpu.us-central1-a.c.ollama.internal ollama[5042]: {"function":"launch_slot_with_data","level":"INFO","line":804,"msg":"slot is processing task","slot_id":0,"task_id":29930,"tid":"140079034640064","timestamp":1712207704} Apr 04 05:15:04 gpu.us-central1-a.c.ollama.internal ollama[5042]: {"function":"update_slots","level":"INFO","line":1808,"msg":"kv cache rm [p0, end)","p0":0,"slot_id":0,"task_id":29930,"tid":"140079034640064","timestamp":1712207704} Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 256 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 128 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 64 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 32 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 16 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 8 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 4 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 2 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 1 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to decode the batch, n_batch = 1, ret = 1 Apr 04 05:15:44 gpu.us-central1-a.c.ollama.internal ollama[5042]: [1712207744] update_slots : failed to find free space in the KV cache, retrying with smaller n_batch = 256 ``` ### What did you expect to see? _No response_ ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS _No response_ ### Architecture _No response_ ### Platform _No response_ ### Ollama version _No response_ ### GPU _No response_ ### GPU info _No response_ ### CPU _No response_ ### Other software _No response_
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3483/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3483/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/167
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/167/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/167/comments
https://api.github.com/repos/ollama/ollama/issues/167/events
https://github.com/ollama/ollama/pull/167
1,816,432,518
PR_kwDOJ0Z1Ps5WIaj5
167
partial decode ggml bin for more info
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-07-21T22:10:33
2023-08-11T00:22:41
2023-08-11T00:22:40
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/167", "html_url": "https://github.com/ollama/ollama/pull/167", "diff_url": "https://github.com/ollama/ollama/pull/167.diff", "patch_url": "https://github.com/ollama/ollama/pull/167.patch", "merged_at": "2023-08-11T00:22:40" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/167/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/167/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/27
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/27/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/27/comments
https://api.github.com/repos/ollama/ollama/issues/27/events
https://github.com/ollama/ollama/issues/27
1,782,711,141
I_kwDOJ0Z1Ps5qQgNl
27
show a loading bar for model loading
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5667396205, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2abQ", "url": "https://api.github.com/repos/ollama/ollama/labels/help%20wanted", "name": "help wanted", "color": "008672", "default": true, "description": "Extra attention is needed" } ]
closed
false
null
[]
null
0
2023-06-30T15:17:59
2023-09-07T11:11:10
2023-09-07T11:11:10
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
currently we show a spinner – even better would be to show a loader similar to the download one
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/27/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/27/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4749
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4749/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4749/comments
https://api.github.com/repos/ollama/ollama/issues/4749/events
https://github.com/ollama/ollama/issues/4749
2,327,508,143
I_kwDOJ0Z1Ps6KuvSv
4,749
OLLAMA_MODELS not applied on initial start or on restart after upgrade on macOS
{ "login": "vernonstinebaker", "id": 8304954, "node_id": "MDQ6VXNlcjgzMDQ5NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8304954?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vernonstinebaker", "html_url": "https://github.com/vernonstinebaker", "followers_url": "https://api.github.com/users/vernonstinebaker/followers", "following_url": "https://api.github.com/users/vernonstinebaker/following{/other_user}", "gists_url": "https://api.github.com/users/vernonstinebaker/gists{/gist_id}", "starred_url": "https://api.github.com/users/vernonstinebaker/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vernonstinebaker/subscriptions", "organizations_url": "https://api.github.com/users/vernonstinebaker/orgs", "repos_url": "https://api.github.com/users/vernonstinebaker/repos", "events_url": "https://api.github.com/users/vernonstinebaker/events{/privacy}", "received_events_url": "https://api.github.com/users/vernonstinebaker/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
4
2024-05-31T10:32:39
2024-07-26T03:15:28
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
The addition of OLLAMA_MODELS is much appreciated, since it allows specifying a different location, such as an external disk, where more space might be available. One issue, however, is if we put specify OLLAMA_MODELS in our .zshrc, for example, the .zshrc file isn't read when Ollama starts initially or when Ollama restarts after an update. Perhaps I'm missing something? Otherwise, it would be great to make this something that can be configured/set directly in Ollama instead of needing to quit Ollama, open a Terminal (so that .zshrc is read) and start Ollama from the Terminal so that the OLLAMA_MODELS directory is set instead of using the default.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4749/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4749/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6280
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6280/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6280/comments
https://api.github.com/repos/ollama/ollama/issues/6280/events
https://github.com/ollama/ollama/issues/6280
2,457,405,573
I_kwDOJ0Z1Ps6SeQiF
6,280
Need qwen2:math !!
{ "login": "jsrdcht", "id": 22300065, "node_id": "MDQ6VXNlcjIyMzAwMDY1", "avatar_url": "https://avatars.githubusercontent.com/u/22300065?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jsrdcht", "html_url": "https://github.com/jsrdcht", "followers_url": "https://api.github.com/users/jsrdcht/followers", "following_url": "https://api.github.com/users/jsrdcht/following{/other_user}", "gists_url": "https://api.github.com/users/jsrdcht/gists{/gist_id}", "starred_url": "https://api.github.com/users/jsrdcht/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jsrdcht/subscriptions", "organizations_url": "https://api.github.com/users/jsrdcht/orgs", "repos_url": "https://api.github.com/users/jsrdcht/repos", "events_url": "https://api.github.com/users/jsrdcht/events{/privacy}", "received_events_url": "https://api.github.com/users/jsrdcht/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-08-09T08:32:20
2024-09-02T23:12:26
2024-09-02T23:12:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://github.com/QwenLM/Qwen2-Math
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6280/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6280/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8303
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8303/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8303/comments
https://api.github.com/repos/ollama/ollama/issues/8303/events
https://github.com/ollama/ollama/issues/8303
2,768,632,208
I_kwDOJ0Z1Ps6lBfmQ
8,303
Warning: could not connect to a running Ollama instance
{ "login": "rcjdut", "id": 140966542, "node_id": "U_kgDOCGb6jg", "avatar_url": "https://avatars.githubusercontent.com/u/140966542?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rcjdut", "html_url": "https://github.com/rcjdut", "followers_url": "https://api.github.com/users/rcjdut/followers", "following_url": "https://api.github.com/users/rcjdut/following{/other_user}", "gists_url": "https://api.github.com/users/rcjdut/gists{/gist_id}", "starred_url": "https://api.github.com/users/rcjdut/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rcjdut/subscriptions", "organizations_url": "https://api.github.com/users/rcjdut/orgs", "repos_url": "https://api.github.com/users/rcjdut/repos", "events_url": "https://api.github.com/users/rcjdut/events{/privacy}", "received_events_url": "https://api.github.com/users/rcjdut/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2025-01-04T08:22:23
2025-01-04T08:40:22
2025-01-04T08:40:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I manually installed Ollama and successfully ran the command 'ollama serve' in a terminal,but when i try to ran the command 'ollama -v',it showed 'Warning: could not connect to a running Ollama instance Warning: client version is 0.5.4' ![屏幕截图 2025-01-04 162123](https://github.com/user-attachments/assets/124d01b6-d540-4d89-8b4f-b84e59cdc274) ![屏幕截图 2025-01-04 162203](https://github.com/user-attachments/assets/c2815813-408d-46c6-a4ad-16530c4c7e50) ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version _No response_
{ "login": "rcjdut", "id": 140966542, "node_id": "U_kgDOCGb6jg", "avatar_url": "https://avatars.githubusercontent.com/u/140966542?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rcjdut", "html_url": "https://github.com/rcjdut", "followers_url": "https://api.github.com/users/rcjdut/followers", "following_url": "https://api.github.com/users/rcjdut/following{/other_user}", "gists_url": "https://api.github.com/users/rcjdut/gists{/gist_id}", "starred_url": "https://api.github.com/users/rcjdut/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rcjdut/subscriptions", "organizations_url": "https://api.github.com/users/rcjdut/orgs", "repos_url": "https://api.github.com/users/rcjdut/repos", "events_url": "https://api.github.com/users/rcjdut/events{/privacy}", "received_events_url": "https://api.github.com/users/rcjdut/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8303/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8303/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/692
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/692/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/692/comments
https://api.github.com/repos/ollama/ollama/issues/692/events
https://github.com/ollama/ollama/pull/692
1,924,891,665
PR_kwDOJ0Z1Ps5b1LE3
692
Increase client/server streaming buffer size to prevent `token too long` error
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-10-03T20:28:42
2023-10-04T18:09:01
2023-10-04T18:09:00
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/692", "html_url": "https://github.com/ollama/ollama/pull/692", "diff_url": "https://github.com/ollama/ollama/pull/692.diff", "patch_url": "https://github.com/ollama/ollama/pull/692.patch", "merged_at": "2023-10-04T18:09:00" }
In the case of a large input the response from `/generate` would be very long due to the encoded context length. Increase the buffer size to prevent this error. resolves #687
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/692/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/692/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1224
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1224/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1224/comments
https://api.github.com/repos/ollama/ollama/issues/1224/events
https://github.com/ollama/ollama/pull/1224
2,004,834,250
PR_kwDOJ0Z1Ps5gDPzg
1,224
update llama.cpp
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-11-21T17:50:16
2023-11-21T20:22:00
2023-11-21T20:21:59
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1224", "html_url": "https://github.com/ollama/ollama/pull/1224", "diff_url": "https://github.com/ollama/ollama/pull/1224.diff", "patch_url": "https://github.com/ollama/ollama/pull/1224.patch", "merged_at": "2023-11-21T20:21:59" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1224/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1224/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5129
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5129/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5129/comments
https://api.github.com/repos/ollama/ollama/issues/5129/events
https://github.com/ollama/ollama/issues/5129
2,361,138,517
I_kwDOJ0Z1Ps6MvB1V
5,129
gpu destroys TMPDIR
{ "login": "vt-alt", "id": 36664211, "node_id": "MDQ6VXNlcjM2NjY0MjEx", "avatar_url": "https://avatars.githubusercontent.com/u/36664211?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vt-alt", "html_url": "https://github.com/vt-alt", "followers_url": "https://api.github.com/users/vt-alt/followers", "following_url": "https://api.github.com/users/vt-alt/following{/other_user}", "gists_url": "https://api.github.com/users/vt-alt/gists{/gist_id}", "starred_url": "https://api.github.com/users/vt-alt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vt-alt/subscriptions", "organizations_url": "https://api.github.com/users/vt-alt/orgs", "repos_url": "https://api.github.com/users/vt-alt/repos", "events_url": "https://api.github.com/users/vt-alt/events{/privacy}", "received_events_url": "https://api.github.com/users/vt-alt/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-06-19T02:53:24
2024-06-25T07:16:32
2024-06-20T17:41:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When building ollama package for ALT Linux I noticed that %buildroot (directory where new binaries are installed) is disappeared after `go test github.com/ollama/ollama/gpu`. Our `%buildroot` (`/usr/src/tmp/ollama-buildroot`) is inside of `TMPDIR` (`/usr/src/tmp`). I am not completely investigated the issue but this is extremely dangerous practice to delete directories you did not create. If it deletes %buildroot it may delete anything else such as homedir? I see so suspicious code fragment in `gpu/assets.go`: ```go func Cleanup() { lock.Lock() defer lock.Unlock() runnersDir := envconfig.RunnersDir if payloadsDir != "" && runnersDir == "" && runtime.GOOS != "windows" { // We want to fully clean up the tmpdir parent of the payloads dir tmpDir := filepath.Clean(filepath.Join(payloadsDir, "..")) slog.Debug("cleaning up", "dir", tmpDir) err := os.RemoveAll(tmpDir) if err != nil { // On windows, if we remove too quickly the llama.dll may still be in-use and fail to remove time.Sleep(1000 * time.Millisecond) err = os.RemoveAll(tmpDir) if err != nil { slog.Warn("failed to clean up", "dir", tmpDir, "err", err) } } } } ``` So it tries to delete parent of TMPDIR? What if HOME or other important directories are there? Please make it not delete directories or files it did not create ### OS Linux ### GPU Other ### CPU Intel ### Ollama version 0.1.44
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5129/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5129/timeline
null
completed
false