url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/3963
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3963/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3963/comments
https://api.github.com/repos/ollama/ollama/issues/3963/events
https://github.com/ollama/ollama/pull/3963
2,266,554,114
PR_kwDOJ0Z1Ps5t4mC3
3,963
windows: init `cmakeTargets` in `init_vars`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-04-26T21:57:05
2024-04-26T22:40:31
2024-04-26T22:40:31
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3963", "html_url": "https://github.com/ollama/ollama/pull/3963", "diff_url": "https://github.com/ollama/ollama/pull/3963.diff", "patch_url": "https://github.com/ollama/ollama/pull/3963.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3963/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3963/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/952
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/952/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/952/comments
https://api.github.com/repos/ollama/ollama/issues/952/events
https://github.com/ollama/ollama/pull/952
1,969,294,048
PR_kwDOJ0Z1Ps5eK5xF
952
support raw generation requests
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-10-30T22:01:26
2023-11-08T22:05:03
2023-11-08T22:05:02
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/952", "html_url": "https://github.com/ollama/ollama/pull/952", "diff_url": "https://github.com/ollama/ollama/pull/952.diff", "patch_url": "https://github.com/ollama/ollama/pull/952.patch", "merged_at": "2023-11-08T22:05:02" }
- add the optional `raw` generate request parameter to bypass prompt formatting and response context Add a `raw` parameter to `/generate` requests that allow directly specifying the prompt without the Ollama server applying additional formatting. ```bash curl -X "POST" -d '{"model":"mistral", "prompt": "[INST] hi [/INST]", "raw": true, "stream": false}' 'http://127.0.0.1:11434/api/generate' ``` Example use case, few-shot prompting: ```python import requests def call_generate_endpoint(prompt, model="mistral", raw=True, stream=False): url = "http://127.0.0.1:11434/api/generate" formatted_prompt = f"""[INST] This is awesome! [/INST] Postive [INST] This is bad! [/INST] Negative [INST] I love this movie [/INST] Positive [INST] {prompt} [/INST] """ payload = { "model": model, "prompt": formatted_prompt, "raw": raw, "stream": stream } response = requests.post(url, json=payload) return response.json() resp = call_generate_endpoint("I hate this book") print(resp.response) # Negative ```
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/952/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/952/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1082
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1082/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1082/comments
https://api.github.com/repos/ollama/ollama/issues/1082/events
https://github.com/ollama/ollama/pull/1082
1,988,676,727
PR_kwDOJ0Z1Ps5fMqVc
1,082
Add support for Multimodel models
{ "login": "mattapperson", "id": 192433, "node_id": "MDQ6VXNlcjE5MjQzMw==", "avatar_url": "https://avatars.githubusercontent.com/u/192433?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mattapperson", "html_url": "https://github.com/mattapperson", "followers_url": "https://api.github.com/users/mattapperson/followers", "following_url": "https://api.github.com/users/mattapperson/following{/other_user}", "gists_url": "https://api.github.com/users/mattapperson/gists{/gist_id}", "starred_url": "https://api.github.com/users/mattapperson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mattapperson/subscriptions", "organizations_url": "https://api.github.com/users/mattapperson/orgs", "repos_url": "https://api.github.com/users/mattapperson/repos", "events_url": "https://api.github.com/users/mattapperson/events{/privacy}", "received_events_url": "https://api.github.com/users/mattapperson/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
12
2023-11-11T02:57:28
2023-11-22T00:22:46
2023-11-22T00:22:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1082", "html_url": "https://github.com/ollama/ollama/pull/1082", "diff_url": "https://github.com/ollama/ollama/pull/1082.diff", "patch_url": "https://github.com/ollama/ollama/pull/1082.patch", "merged_at": null }
Interactive cli usage: ```bash /set image add <image id int> <path to image file> Please tell me what text is in this photo [img-<image id int>] ``` For the API I added support for the `image_data` prop with the type of `[]{id: int, data: string(base64)}`for the generate endpoint. To support this, modelfile now has a `MMPROJ` key that points to the mmproj file path. Tested with the following modelfile: ``` FROM ./ggml-model-q4_k.gguf TEMPLATE """ USER:{{ .Prompt }} ASSISTANT: """ MMPROJ ./mmproj-model-f16.gguf ``` And using the following pre-quantatized model - https://huggingface.co/mys/ggml_llava-v1.5-13b. <img width="1226" alt="Screenshot 2023-11-09 at 11 30 10 PM" src="https://github.com/jmorganca/ollama/assets/192433/ec62e5e7-999b-44f2-9a32-e6b85ca5b404"> <img width="567" alt="screenshot" src="https://github.com/jmorganca/ollama/assets/192433/681c5cd5-9140-43b1-8383-08823cdbb002">
{ "login": "mattapperson", "id": 192433, "node_id": "MDQ6VXNlcjE5MjQzMw==", "avatar_url": "https://avatars.githubusercontent.com/u/192433?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mattapperson", "html_url": "https://github.com/mattapperson", "followers_url": "https://api.github.com/users/mattapperson/followers", "following_url": "https://api.github.com/users/mattapperson/following{/other_user}", "gists_url": "https://api.github.com/users/mattapperson/gists{/gist_id}", "starred_url": "https://api.github.com/users/mattapperson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mattapperson/subscriptions", "organizations_url": "https://api.github.com/users/mattapperson/orgs", "repos_url": "https://api.github.com/users/mattapperson/repos", "events_url": "https://api.github.com/users/mattapperson/events{/privacy}", "received_events_url": "https://api.github.com/users/mattapperson/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1082/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1082/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8141
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8141/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8141/comments
https://api.github.com/repos/ollama/ollama/issues/8141/events
https://github.com/ollama/ollama/issues/8141
2,745,757,781
I_kwDOJ0Z1Ps6jqPBV
8,141
command r plus model update
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/followers", "following_url": "https://api.github.com/users/olumolu/following{/other_user}", "gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}", "starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/olumolu/subscriptions", "organizations_url": "https://api.github.com/users/olumolu/orgs", "repos_url": "https://api.github.com/users/olumolu/repos", "events_url": "https://api.github.com/users/olumolu/events{/privacy}", "received_events_url": "https://api.github.com/users/olumolu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
1
2024-12-17T18:48:33
2024-12-23T08:13:29
2024-12-23T08:13:29
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
> https://docs.cohere.com/docs/command-r-plus They already have updated there model can we have the updated new model in olllama thanks
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8141/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8141/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/5864
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5864/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5864/comments
https://api.github.com/repos/ollama/ollama/issues/5864/events
https://github.com/ollama/ollama/pull/5864
2,423,919,879
PR_kwDOJ0Z1Ps52Je6A
5,864
Bump Go patch version
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-22T23:17:38
2024-07-22T23:34:22
2024-07-22T23:34:19
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5864", "html_url": "https://github.com/ollama/ollama/pull/5864", "diff_url": "https://github.com/ollama/ollama/pull/5864.diff", "patch_url": "https://github.com/ollama/ollama/pull/5864.patch", "merged_at": "2024-07-22T23:34:18" }
Fixes #5774
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5864/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5864/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/321
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/321/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/321/comments
https://api.github.com/repos/ollama/ollama/issues/321/events
https://github.com/ollama/ollama/pull/321
1,846,003,602
PR_kwDOJ0Z1Ps5XruCs
321
length check for parameters
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-08-10T23:09:25
2023-08-10T23:23:11
2023-08-10T23:23:10
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/321", "html_url": "https://github.com/ollama/ollama/pull/321", "diff_url": "https://github.com/ollama/ollama/pull/321.diff", "patch_url": "https://github.com/ollama/ollama/pull/321.patch", "merged_at": "2023-08-10T23:23:10" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/321/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/321/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1392
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1392/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1392/comments
https://api.github.com/repos/ollama/ollama/issues/1392/events
https://github.com/ollama/ollama/pull/1392
2,026,923,082
PR_kwDOJ0Z1Ps5hN7DP
1,392
chat api endpoint
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-12-05T18:57:45
2023-12-11T07:33:21
2023-12-05T19:57:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1392", "html_url": "https://github.com/ollama/ollama/pull/1392", "diff_url": "https://github.com/ollama/ollama/pull/1392.diff", "patch_url": "https://github.com/ollama/ollama/pull/1392.patch", "merged_at": "2023-12-05T19:57:33" }
- add a new `/api/chat` API endpoint that takes an array of `message` objects. This endpoint is an alternative to `/api/generate`. - deprecate generation context and template, but continue to support them - rebuild chat content from messages This changes adds a `/api/chat` endpoint to the API which takes an array of messages. This makes modifying and tracking the history on the fly much simpler. It is an alternative to prompt/response. `context` will continue to work as expected for now, but at some point in the future we may want to replace it completely with `/api/chat`. ``` curl -X POST http://localhost:11434/api/generate -d '{ "model": "mistral", "prompt": "hello, how are you?" }' OR ### Basic generate request with messages curl -X POST http://localhost:11434/api/chat -d '{ "model": "mistral", "messages": [ { "role": "user", "content": "why is the sky blue?" } ] }' ``` resolves #981 resolves #1203
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1392/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1392/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1940
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1940/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1940/comments
https://api.github.com/repos/ollama/ollama/issues/1940/events
https://github.com/ollama/ollama/issues/1940
2,077,872,259
I_kwDOJ0Z1Ps572dCD
1,940
`CUDA error 100` after detecting GPU libraries on system
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
10
2024-01-12T01:13:13
2024-02-19T19:50:24
2024-02-19T19:50:24
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It seems that upon detecting an Nvidia card, `ollama` may error with `CUDA error 100`: ``` Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:88: Detecting GPU type Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:203: Searching for GPU management library libnvidia-ml.so Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:248: Discovered GPU libraries: [/usr/lib/x86_64-linux-gnu/libnvidia-ml.so.535.104.05 /usr/lib/wsl/lib/libnvidia-ml.so.1] Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:259: Unable to load CUDA management library /usr/lib/x86_64-linux-gnu/libnvidia-ml.so.535.104.05: nvml vram init failure: 9 Jan 11 15:37:51 LR9135SQP ollama[5616]: 2024/01/11 15:37:51 gpu.go:94: Nvidia GPU detected Jan 11 15:37:51 LR9135SQP ollama[5616]: 2024/01/11 15:37:51 gpu.go:135: CUDA Compute Capability detected: 7.5 ``` ``` Jan 11 15:55:41 LR9135SQP ollama[5616]: CUDA error 100 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/ggml-cuda.cu:495: no CUDA-capable device is detected Jan 11 15:55:41 LR9135SQP ollama[5616]: current device: 1881676272 Jan 11 15:55:41 LR9135SQP ollama[5616]: Lazy loading /tmp/ollama958766944/cuda/libext_server.so library Jan 11 15:55:41 LR9135SQP ollama[5616]: GGML_ASSERT: /go/src/github.com/jmorganca/ollama/llm/llama.cpp/ggml-cuda.cu:495: !"CUDA error" ```
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1940/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1940/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6839
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6839/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6839/comments
https://api.github.com/repos/ollama/ollama/issues/6839/events
https://github.com/ollama/ollama/issues/6839
2,531,251,465
I_kwDOJ0Z1Ps6W39UJ
6,839
ollama request llama3.1 fail.
{ "login": "microbitcswcss", "id": 129246500, "node_id": "U_kgDOB7QlJA", "avatar_url": "https://avatars.githubusercontent.com/u/129246500?v=4", "gravatar_id": "", "url": "https://api.github.com/users/microbitcswcss", "html_url": "https://github.com/microbitcswcss", "followers_url": "https://api.github.com/users/microbitcswcss/followers", "following_url": "https://api.github.com/users/microbitcswcss/following{/other_user}", "gists_url": "https://api.github.com/users/microbitcswcss/gists{/gist_id}", "starred_url": "https://api.github.com/users/microbitcswcss/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/microbitcswcss/subscriptions", "organizations_url": "https://api.github.com/users/microbitcswcss/orgs", "repos_url": "https://api.github.com/users/microbitcswcss/repos", "events_url": "https://api.github.com/users/microbitcswcss/events{/privacy}", "received_events_url": "https://api.github.com/users/microbitcswcss/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
5
2024-09-17T14:00:01
2024-10-28T07:39:53
2024-10-28T07:39:52
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? like title i cant request llama3.1 on windows 10 it can be used one month ago, cmd.exe: C:\Users\123>ollama run llama3.1 2024/09/17 21:53:18 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434 2024/09/17 21:53:18 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434 Error: Head "http://C:11434/": dial tcp: lookup C: no such host C:\Users\123 >ollama --version 2024/09/17 21:59:13 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434 2024/09/17 21:59:13 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434 Warning: could not connect to a running Ollama instance Warning: client version is 0.3.9 ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.3.9
{ "login": "microbitcswcss", "id": 129246500, "node_id": "U_kgDOB7QlJA", "avatar_url": "https://avatars.githubusercontent.com/u/129246500?v=4", "gravatar_id": "", "url": "https://api.github.com/users/microbitcswcss", "html_url": "https://github.com/microbitcswcss", "followers_url": "https://api.github.com/users/microbitcswcss/followers", "following_url": "https://api.github.com/users/microbitcswcss/following{/other_user}", "gists_url": "https://api.github.com/users/microbitcswcss/gists{/gist_id}", "starred_url": "https://api.github.com/users/microbitcswcss/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/microbitcswcss/subscriptions", "organizations_url": "https://api.github.com/users/microbitcswcss/orgs", "repos_url": "https://api.github.com/users/microbitcswcss/repos", "events_url": "https://api.github.com/users/microbitcswcss/events{/privacy}", "received_events_url": "https://api.github.com/users/microbitcswcss/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6839/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6839/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/220
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/220/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/220/comments
https://api.github.com/repos/ollama/ollama/issues/220/events
https://github.com/ollama/ollama/issues/220
1,822,783,975
I_kwDOJ0Z1Ps5spXnn
220
Sending input with a + crashes
{ "login": "benr75", "id": 661, "node_id": "MDQ6VXNlcjY2MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/661?v=4", "gravatar_id": "", "url": "https://api.github.com/users/benr75", "html_url": "https://github.com/benr75", "followers_url": "https://api.github.com/users/benr75/followers", "following_url": "https://api.github.com/users/benr75/following{/other_user}", "gists_url": "https://api.github.com/users/benr75/gists{/gist_id}", "starred_url": "https://api.github.com/users/benr75/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/benr75/subscriptions", "organizations_url": "https://api.github.com/users/benr75/orgs", "repos_url": "https://api.github.com/users/benr75/repos", "events_url": "https://api.github.com/users/benr75/events{/privacy}", "received_events_url": "https://api.github.com/users/benr75/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2023-07-26T16:37:12
2023-08-04T14:28:22
2023-08-04T14:28:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Maybe an url escaping issue? I didn't dig much deeper but wanted to report. ``` $ ollama run llama2:13b >>> I cannot send input with a + in it Error: Post "http://127.0.0.1:11434/api/generate": EOF ```
{ "login": "benr75", "id": 661, "node_id": "MDQ6VXNlcjY2MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/661?v=4", "gravatar_id": "", "url": "https://api.github.com/users/benr75", "html_url": "https://github.com/benr75", "followers_url": "https://api.github.com/users/benr75/followers", "following_url": "https://api.github.com/users/benr75/following{/other_user}", "gists_url": "https://api.github.com/users/benr75/gists{/gist_id}", "starred_url": "https://api.github.com/users/benr75/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/benr75/subscriptions", "organizations_url": "https://api.github.com/users/benr75/orgs", "repos_url": "https://api.github.com/users/benr75/repos", "events_url": "https://api.github.com/users/benr75/events{/privacy}", "received_events_url": "https://api.github.com/users/benr75/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/220/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/220/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8248
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8248/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8248/comments
https://api.github.com/repos/ollama/ollama/issues/8248/events
https://github.com/ollama/ollama/issues/8248
2,759,846,975
I_kwDOJ0Z1Ps6kf-w_
8,248
Unexpected Connection Closure and GPU Memory Not Releasing
{ "login": "BronyaKaslana06", "id": 97924591, "node_id": "U_kgDOBdY17w", "avatar_url": "https://avatars.githubusercontent.com/u/97924591?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BronyaKaslana06", "html_url": "https://github.com/BronyaKaslana06", "followers_url": "https://api.github.com/users/BronyaKaslana06/followers", "following_url": "https://api.github.com/users/BronyaKaslana06/following{/other_user}", "gists_url": "https://api.github.com/users/BronyaKaslana06/gists{/gist_id}", "starred_url": "https://api.github.com/users/BronyaKaslana06/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BronyaKaslana06/subscriptions", "organizations_url": "https://api.github.com/users/BronyaKaslana06/orgs", "repos_url": "https://api.github.com/users/BronyaKaslana06/repos", "events_url": "https://api.github.com/users/BronyaKaslana06/events{/privacy}", "received_events_url": "https://api.github.com/users/BronyaKaslana06/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
7
2024-12-26T14:40:51
2025-01-01T02:27:55
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? ### Problem Description I am using Ubuntu 22.04 and making network requests to a local Ollama service with Python to run a series of models sequentially. After each model runs, it is unloaded using the following Python code. However, the task encounters an issue after reaching a certain point, where the Ollama network service unexpectedly stops when loading a model, resulting in the following error: ``` Traceback (most recent call last): File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 703, in urlopen httplib_response = self._make_request( File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 449, in _make_request six.raise_from(e, None) File "<string>", line 3, in raise_from File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 444, in _make_request httplib_response = conn.getresponse() File "/root/miniconda3/lib/python3.10/http/client.py", line 1374, in getresponse response.begin() File "/root/miniconda3/lib/python3.10/http/client.py", line 318, in begin version, status, reason = self._read_status() File "/root/miniconda3/lib/python3.10/http/client.py", line 287, in _read_status raise RemoteDisconnected("Remote end closed connection without" http.client.RemoteDisconnected: Remote end closed connection without response During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/root/miniconda3/lib/python3.10/site-packages/requests/adapters.py", line 667, in send resp = conn.urlopen( File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 787, in urlopen retries = retries.increment( File "/root/miniconda3/lib/python3.10/site-packages/urllib3/util/retry.py", line 550, in increment raise six.reraise(type(error), error, _stacktrace) File "/root/miniconda3/lib/python3.10/site-packages/urllib3/packages/six.py", line 769, in reraise raise value.with_traceback(tb) File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 703, in urlopen httplib_response = self._make_request( File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 449, in _make_request six.raise_from(e, None) File "<string>", line 3, in raise_from File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 444, in _make_request httplib_response = conn.getresponse() File "/root/miniconda3/lib/python3.10/http/client.py", line 1374, in getresponse response.begin() File "/root/miniconda3/lib/python3.10/http/client.py", line 318, in begin version, status, reason = self._read_status() File "/root/miniconda3/lib/python3.10/http/client.py", line 287, in _read_status raise RemoteDisconnected("Remote end closed connection without" urllib3.exceptions.ProtocolError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/root/autodl-tmp/EvalLLM/main.py", line 109, in <module> generator.generate_and_save(section) File "/root/autodl-tmp/EvalLLM/main.py", line 93, in generate_and_save response_text = ollamaCaller.generate_response(prompt=question) File "/root/autodl-tmp/EvalLLM/Caller/OllamaCaller.py", line 18, in generate_response response = requests.post("http://localhost:11434/api/generate", json=payload) File "/root/miniconda3/lib/python3.10/site-packages/requests/api.py", line 115, in post return request("post", url, data=data, json=json, **kwargs) File "/root/miniconda3/lib/python3.10/site-packages/requests/api.py", line 59, in request return session.request(method=method, url=url, **kwargs) File "/root/miniconda3/lib/python3.10/site-packages/requests/sessions.py", line 589, in request resp = self.send(prep, **send_kwargs) File "/root/miniconda3/lib/python3.10/site-packages/requests/sessions.py", line 703, in send r = adapter.send(request, **kwargs) File "/root/miniconda3/lib/python3.10/site-packages/requests/adapters.py", line 682, in send raise ConnectionError(err, request=request) requests.exceptions.ConnectionError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response')) ``` This could be caused by the **service not responding for a long time**, or it might be due to the **daemon process being unexpectedly closed**, right? At the time of the error, The program was just finished loading a new model (**a relatively small model for my gpu, llama3:latest**), and the Ollama logs are as follows: ```log time=2024-12-25T22:05:34.725+08:00 level=INFO source=sched.go:714 msg="new model will fit in available VRAM in single GPU, loading" model=/root/autodl-tmp/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa gpu=GPU-30d02008-3980-d577-cd07-7e3fcabf1d31 parallel=4 available=25158156288 required="6.2 GiB" time=2024-12-25T22:05:34.977+08:00 level=INFO source=server.go:104 msg="system memory" total="377.3 GiB" free="327.5 GiB" free_swap="0 B" time=2024-12-25T22:05:34.977+08:00 level=INFO source=memory.go:356 msg="offload to cuda" layers.requested=-1 layers.model=33 layers.offload=33 layers.split="" memory.available="[23.4 GiB]" memory.gpu_overhead="0 B" memory.required.full="6.2 GiB" memory.required.partial="6.2 GiB" memory.required.kv="1.0 GiB" memory.required.allocations="[6.2 GiB]" memory.weights.total="4.7 GiB" memory.weights.repeating="4.3 GiB" memory.weights.nonrepeating="411.0 MiB" memory.graph.full="560.0 MiB" memory.graph.partial="677.5 MiB" time=2024-12-25T22:05:34.977+08:00 level=INFO source=server.go:376 msg="starting llama server" cmd="/usr/local/lib/ollama/runners/cuda_v12_avx/ollama_llama_server runner --model /root/autodl-tmp/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa --ctx-size 8192 --batch-size 512 --n-gpu-layers 33 --threads 112 --parallel 4 --port 41461" time=2024-12-25T22:05:34.978+08:00 level=INFO source=sched.go:449 msg="loaded runners" count=1 time=2024-12-25T22:05:34.978+08:00 level=INFO source=server.go:555 msg="waiting for llama runner to start responding" time=2024-12-25T22:05:34.978+08:00 level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server error" time=2024-12-25T22:05:35.021+08:00 level=INFO source=runner.go:945 msg="starting go runner" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes time=2024-12-25T22:05:35.029+08:00 level=INFO source=runner.go:946 msg=system info="CUDA : ARCHS = 600,610,620,700,720,750,800,860,870,890,900 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | LLAMAFILE = 1 | AARCH64_REPACK = 1 | cgo(gcc)" threads=112 time=2024-12-25T22:05:35.029+08:00 level=INFO source=.:0 msg="Server listening on 127.0.0.1:41461" llama_load_model_from_file: using device CUDA0 (NVIDIA GeForce RTX 3090) - 23992 MiB free llama_model_loader: loaded meta data with 22 key-value pairs and 291 tensors from /root/autodl-tmp/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = llama llama_model_loader: - kv 1: general.name str = Meta-Llama-3-8B-Instruct llama_model_loader: - kv 2: llama.block_count u32 = 32 llama_model_loader: - kv 3: llama.context_length u32 = 8192 llama_model_loader: - kv 4: llama.embedding_length u32 = 4096 llama_model_loader: - kv 5: llama.feed_forward_length u32 = 14336 llama_model_loader: - kv 6: llama.attention.head_count u32 = 32 llama_model_loader: - kv 7: llama.attention.head_count_kv u32 = 8 llama_model_loader: - kv 8: llama.rope.freq_base f32 = 500000.000000 llama_model_loader: - kv 9: llama.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 10: general.file_type u32 = 2 llama_model_loader: - kv 11: llama.vocab_size u32 = 128256 llama_model_loader: - kv 12: llama.rope.dimension_count u32 = 128 llama_model_loader: - kv 13: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 14: tokenizer.ggml.pre str = llama-bpe llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,128256] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 16: tokenizer.ggml.token_type arr[i32,128256] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... time=2024-12-25T22:05:35.230+08:00 level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server loading model" llama_model_loader: - kv 17: tokenizer.ggml.merges arr[str,280147] = ["Ġ Ġ", "Ġ ĠĠĠ", "ĠĠ ĠĠ", "... llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 128000 llama_model_loader: - kv 19: tokenizer.ggml.eos_token_id u32 = 128009 llama_model_loader: - kv 20: tokenizer.chat_template str = {% set loop_messages = messages %}{% ... llama_model_loader: - kv 21: general.quantization_version u32 = 2 llama_model_loader: - type f32: 65 tensors llama_model_loader: - type q4_0: 225 tensors llama_model_loader: - type q6_K: 1 tensors llm_load_vocab: special tokens cache size = 256 llm_load_vocab: token to piece cache size = 0.8000 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = llama llm_load_print_meta: vocab type = BPE llm_load_print_meta: n_vocab = 128256 llm_load_print_meta: n_merges = 280147 llm_load_print_meta: vocab_only = 0 llm_load_print_meta: n_ctx_train = 8192 llm_load_print_meta: n_embd = 4096 llm_load_print_meta: n_layer = 32 llm_load_print_meta: n_head = 32 llm_load_print_meta: n_head_kv = 8 llm_load_print_meta: n_rot = 128 llm_load_print_meta: n_swa = 0 llm_load_print_meta: n_embd_head_k = 128 llm_load_print_meta: n_embd_head_v = 128 llm_load_print_meta: n_gqa = 4 llm_load_print_meta: n_embd_k_gqa = 1024 llm_load_print_meta: n_embd_v_gqa = 1024 llm_load_print_meta: f_norm_eps = 0.0e+00 llm_load_print_meta: f_norm_rms_eps = 1.0e-05 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 14336 llm_load_print_meta: n_expert = 0 llm_load_print_meta: n_expert_used = 0 llm_load_print_meta: causal attn = 1 llm_load_print_meta: pooling type = 0 llm_load_print_meta: rope type = 0 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 500000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_ctx_orig_yarn = 8192 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: ssm_dt_b_c_rms = 0 llm_load_print_meta: model type = 8B llm_load_print_meta: model ftype = Q4_0 llm_load_print_meta: model params = 8.03 B llm_load_print_meta: model size = 4.33 GiB (4.64 BPW) llm_load_print_meta: general.name = Meta-Llama-3-8B-Instruct llm_load_print_meta: BOS token = 128000 '<|begin_of_text|>' llm_load_print_meta: EOS token = 128009 '<|eot_id|>' llm_load_print_meta: EOT token = 128009 '<|eot_id|>' llm_load_print_meta: LF token = 128 'Ä' llm_load_print_meta: EOG token = 128009 '<|eot_id|>' llm_load_print_meta: max token length = 256 llm_load_tensors: offloading 32 repeating layers to GPU llm_load_tensors: offloading output layer to GPU llm_load_tensors: offloaded 33/33 layers to GPU llm_load_tensors: CPU_Mapped model buffer size = 281.81 MiB llm_load_tensors: CUDA0 model buffer size = 4155.99 MiB llama_new_context_with_model: n_seq_max = 4 llama_new_context_with_model: n_ctx = 8192 llama_new_context_with_model: n_ctx_per_seq = 2048 llama_new_context_with_model: n_batch = 2048 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 500000.0 llama_new_context_with_model: freq_scale = 1 llama_new_context_with_model: n_ctx_per_seq (2048) < n_ctx_train (8192) -- the full capacity of the model will not be utilized llama_kv_cache_init: CUDA0 KV buffer size = 1024.00 MiB llama_new_context_with_model: KV self size = 1024.00 MiB, K (f16): 512.00 MiB, V (f16): 512.00 MiB llama_new_context_with_model: CUDA_Host output buffer size = 2.02 MiB llama_new_context_with_model: CUDA0 compute buffer size = 560.00 MiB llama_new_context_with_model: CUDA_Host compute buffer size = 24.01 MiB llama_new_context_with_model: graph nodes = 1030 llama_new_context_with_model: graph splits = 2 ``` The logs end here. Regardless of the cause of this issue, after it occurs, **about 6GB of GPU memory remains unreleased and cannot be freed even by restarting Ollama**. Below is the output of `nvidia-smi`: ``` (base) root@autodl-container-d33848b29e-31b2d2f4:~# nvidia-smi Thu Dec 26 22:29:15 2024 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.78 Driver Version: 550.78 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce RTX 3090 On | 00000000:10:00.0 Off | N/A | | 30% 28C P8 17W / 350W | 6002MiB / 24576MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| +-----------------------------------------------------------------------------------------+ ``` Has anyone encountered this issue? How can it be resolved? Thank you! ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version 0.5.4
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8248/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8248/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7455
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7455/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7455/comments
https://api.github.com/repos/ollama/ollama/issues/7455/events
https://github.com/ollama/ollama/pull/7455
2,627,779,798
PR_kwDOJ0Z1Ps6AkeyU
7,455
Add basic mllama integration tests
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-31T20:37:06
2024-11-01T00:25:51
2024-11-01T00:25:48
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7455", "html_url": "https://github.com/ollama/ollama/pull/7455", "diff_url": "https://github.com/ollama/ollama/pull/7455.diff", "patch_url": "https://github.com/ollama/ollama/pull/7455.patch", "merged_at": "2024-11-01T00:25:48" }
Draft until I run it against a few more systems and make sure the timeout values are reasonable
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7455/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7455/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8379
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8379/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8379/comments
https://api.github.com/repos/ollama/ollama/issues/8379/events
https://github.com/ollama/ollama/issues/8379
2,781,273,717
I_kwDOJ0Z1Ps6lxt51
8,379
Inquiry About Adding vikhyatk/moondream2 to Ollama
{ "login": "Saber120", "id": 108297159, "node_id": "U_kgDOBnR7xw", "avatar_url": "https://avatars.githubusercontent.com/u/108297159?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Saber120", "html_url": "https://github.com/Saber120", "followers_url": "https://api.github.com/users/Saber120/followers", "following_url": "https://api.github.com/users/Saber120/following{/other_user}", "gists_url": "https://api.github.com/users/Saber120/gists{/gist_id}", "starred_url": "https://api.github.com/users/Saber120/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Saber120/subscriptions", "organizations_url": "https://api.github.com/users/Saber120/orgs", "repos_url": "https://api.github.com/users/Saber120/repos", "events_url": "https://api.github.com/users/Saber120/events{/privacy}", "received_events_url": "https://api.github.com/users/Saber120/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2025-01-10T22:03:19
2025-01-13T19:29:39
2025-01-13T19:29:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi Ollama team, I was wondering if you plan to add the model vikhyatk/moondream2, which was just released yesterday, to your platform
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8379/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8379/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1753
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1753/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1753/comments
https://api.github.com/repos/ollama/ollama/issues/1753/events
https://github.com/ollama/ollama/issues/1753
2,061,198,535
I_kwDOJ0Z1Ps5622TH
1,753
[enhancement] Support FROM gpt-4:0314
{ "login": "nathanleclaire", "id": 1476820, "node_id": "MDQ6VXNlcjE0NzY4MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1476820?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nathanleclaire", "html_url": "https://github.com/nathanleclaire", "followers_url": "https://api.github.com/users/nathanleclaire/followers", "following_url": "https://api.github.com/users/nathanleclaire/following{/other_user}", "gists_url": "https://api.github.com/users/nathanleclaire/gists{/gist_id}", "starred_url": "https://api.github.com/users/nathanleclaire/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nathanleclaire/subscriptions", "organizations_url": "https://api.github.com/users/nathanleclaire/orgs", "repos_url": "https://api.github.com/users/nathanleclaire/repos", "events_url": "https://api.github.com/users/nathanleclaire/events{/privacy}", "received_events_url": "https://api.github.com/users/nathanleclaire/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-01-01T00:11:55
2024-01-02T11:26:07
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It would be nice if you could have a `Modelfile` that had an OpenAI model as a base. Then Ollama inference would call the OpenAI APIs instead of local inference with the built parameters. I know it seems like maybe it sorta defeats the point but I really like having the "Ollama facade" available. I use the March version of GPT-4 heavily but have to type "Let's think step by step" and all those other things all the time 😵‍💫 I could look at making a PR for this if there is interest. @jmorganca @mchiang0610
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1753/reactions", "total_count": 5, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 5, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1753/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/226
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/226/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/226/comments
https://api.github.com/repos/ollama/ollama/issues/226/events
https://github.com/ollama/ollama/pull/226
1,824,955,065
PR_kwDOJ0Z1Ps5Wk6Px
226
refactor scan multiline for reuse
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-07-27T18:32:03
2023-07-27T18:45:45
2023-07-27T18:45:41
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/226", "html_url": "https://github.com/ollama/ollama/pull/226", "diff_url": "https://github.com/ollama/ollama/pull/226.diff", "patch_url": "https://github.com/ollama/ollama/pull/226.patch", "merged_at": "2023-07-27T18:45:41" }
It's not obvious values are ingested verbatim when not using multiline so `"` are included in the template. Instead, ingest the value inside the quotes
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/226/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/226/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1866
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1866/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1866/comments
https://api.github.com/repos/ollama/ollama/issues/1866/events
https://github.com/ollama/ollama/issues/1866
2,072,359,110
I_kwDOJ0Z1Ps57hbDG
1,866
Where is the model file stored?
{ "login": "ParisNeo", "id": 827993, "node_id": "MDQ6VXNlcjgyNzk5Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/827993?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParisNeo", "html_url": "https://github.com/ParisNeo", "followers_url": "https://api.github.com/users/ParisNeo/followers", "following_url": "https://api.github.com/users/ParisNeo/following{/other_user}", "gists_url": "https://api.github.com/users/ParisNeo/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParisNeo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParisNeo/subscriptions", "organizations_url": "https://api.github.com/users/ParisNeo/orgs", "repos_url": "https://api.github.com/users/ParisNeo/repos", "events_url": "https://api.github.com/users/ParisNeo/events{/privacy}", "received_events_url": "https://api.github.com/users/ParisNeo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-01-09T13:19:37
2024-01-11T16:49:16
2024-01-09T17:30:51
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi there, I need to make a modification to the model file. Can you please tell me where do you store the model file? Best regards
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1866/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1866/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5531
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5531/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5531/comments
https://api.github.com/repos/ollama/ollama/issues/5531/events
https://github.com/ollama/ollama/issues/5531
2,394,121,820
I_kwDOJ0Z1Ps6Os2Zc
5,531
`ollama create --quantize` does not show proper error if quantizing an unsupported model architecture
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2024-07-07T16:27:14
2024-07-12T00:24:30
2024-07-12T00:24:30
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When quantizing a model with `ollama create -q`, unknown model architecture errors are not properly bubbled up. E.g. from the logs: ``` llama_model_loader: - type f32: 121 tensors llama_model_loader: - type f16: 162 tensors llama_model_quantize: failed to quantize: unknown model architecture: 'chatglm' ``` User error: ``` % ollama create -f Modelfile -q q4_0 test transferring model data quantizing F16 model to Q4_0 Error: llama_model_quantize: 1 ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5531/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5531/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6806
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6806/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6806/comments
https://api.github.com/repos/ollama/ollama/issues/6806/events
https://github.com/ollama/ollama/issues/6806
2,526,589,141
I_kwDOJ0Z1Ps6WmLDV
6,806
slow
{ "login": "ayttop", "id": 178673810, "node_id": "U_kgDOCqZYkg", "avatar_url": "https://avatars.githubusercontent.com/u/178673810?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ayttop", "html_url": "https://github.com/ayttop", "followers_url": "https://api.github.com/users/ayttop/followers", "following_url": "https://api.github.com/users/ayttop/following{/other_user}", "gists_url": "https://api.github.com/users/ayttop/gists{/gist_id}", "starred_url": "https://api.github.com/users/ayttop/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ayttop/subscriptions", "organizations_url": "https://api.github.com/users/ayttop/orgs", "repos_url": "https://api.github.com/users/ayttop/repos", "events_url": "https://api.github.com/users/ayttop/events{/privacy}", "received_events_url": "https://api.github.com/users/ayttop/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
9
2024-09-14T20:03:00
2024-09-25T21:08:43
2024-09-25T21:08:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The program is very slow 3.10 ollama ### OS Windows ### GPU Intel ### CPU Intel ### Ollama version 3.10
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6806/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6806/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1471
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1471/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1471/comments
https://api.github.com/repos/ollama/ollama/issues/1471/events
https://github.com/ollama/ollama/issues/1471
2,036,377,833
I_kwDOJ0Z1Ps55YKjp
1,471
Publishing model fails with "Error: unable to push ..."
{ "login": "stephenwithav", "id": 54563, "node_id": "MDQ6VXNlcjU0NTYz", "avatar_url": "https://avatars.githubusercontent.com/u/54563?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stephenwithav", "html_url": "https://github.com/stephenwithav", "followers_url": "https://api.github.com/users/stephenwithav/followers", "following_url": "https://api.github.com/users/stephenwithav/following{/other_user}", "gists_url": "https://api.github.com/users/stephenwithav/gists{/gist_id}", "starred_url": "https://api.github.com/users/stephenwithav/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stephenwithav/subscriptions", "organizations_url": "https://api.github.com/users/stephenwithav/orgs", "repos_url": "https://api.github.com/users/stephenwithav/repos", "events_url": "https://api.github.com/users/stephenwithav/events{/privacy}", "received_events_url": "https://api.github.com/users/stephenwithav/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2023-12-11T19:25:53
2024-11-03T01:37:33
2024-01-20T00:25:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I followed the [import doc](https://github.com/jmorganca/ollama/blob/7a1b37ac64f0fb0585e279a0a840707843511ed3/docs/import.md?plain=1#L108) steps, but I'm getting the error in the title. I've: 1. Created an account. 2. Copied my `id_ed25519.pub` file to `~/.ollama/` and `/usr/share/ollama/.ollama`. 3. Added the public key to my account. 4. Copied the model to my username's namespace. 5. Attempted `ollama push stephenwithav/book-summary` ``` Error: unable to push stephenwithav/book-summary, make sure this namespace exists and you are authorized to push to it ```
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1471/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1471/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/3923
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3923/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3923/comments
https://api.github.com/repos/ollama/ollama/issues/3923/events
https://github.com/ollama/ollama/pull/3923
2,264,567,354
PR_kwDOJ0Z1Ps5tx0Ne
3,923
precalculate output tensor memory for metal and mmap
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-25T22:00:38
2024-04-25T23:34:18
2024-04-25T23:34:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3923", "html_url": "https://github.com/ollama/ollama/pull/3923", "diff_url": "https://github.com/ollama/ollama/pull/3923.diff", "patch_url": "https://github.com/ollama/ollama/pull/3923.patch", "merged_at": "2024-04-25T23:34:17" }
on metal with mmap, the output tensors are always allocated even if the offloaded layers < total layers + 1. other backends are unaffected
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3923/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3923/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4428
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4428/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4428/comments
https://api.github.com/repos/ollama/ollama/issues/4428/events
https://github.com/ollama/ollama/issues/4428
2,295,167,222
I_kwDOJ0Z1Ps6IzXj2
4,428
Ollama Api not working after adding format:"json"
{ "login": "usmandilmeer", "id": 51738693, "node_id": "MDQ6VXNlcjUxNzM4Njkz", "avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4", "gravatar_id": "", "url": "https://api.github.com/users/usmandilmeer", "html_url": "https://github.com/usmandilmeer", "followers_url": "https://api.github.com/users/usmandilmeer/followers", "following_url": "https://api.github.com/users/usmandilmeer/following{/other_user}", "gists_url": "https://api.github.com/users/usmandilmeer/gists{/gist_id}", "starred_url": "https://api.github.com/users/usmandilmeer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/usmandilmeer/subscriptions", "organizations_url": "https://api.github.com/users/usmandilmeer/orgs", "repos_url": "https://api.github.com/users/usmandilmeer/repos", "events_url": "https://api.github.com/users/usmandilmeer/events{/privacy}", "received_events_url": "https://api.github.com/users/usmandilmeer/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" }, { "id": 6433346500, "node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA", "url": "https://api.github.com/repos/ollama/ollama/labels/amd", "name": "amd", "color": "000000", "default": false, "description": "Issues relating to AMD GPUs and ROCm" } ]
closed
false
null
[]
null
2
2024-05-14T11:23:41
2024-05-14T21:46:32
2024-05-14T21:46:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Ollama api is giving empty response if i add add format as json ![image](https://github.com/ollama/ollama/assets/51738693/31636d22-30bf-4c9f-9a1f-7ac6cc1a33ef) ### OS Windows ### GPU AMD ### CPU Intel ### Ollama version 0.1.37
{ "login": "usmandilmeer", "id": 51738693, "node_id": "MDQ6VXNlcjUxNzM4Njkz", "avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4", "gravatar_id": "", "url": "https://api.github.com/users/usmandilmeer", "html_url": "https://github.com/usmandilmeer", "followers_url": "https://api.github.com/users/usmandilmeer/followers", "following_url": "https://api.github.com/users/usmandilmeer/following{/other_user}", "gists_url": "https://api.github.com/users/usmandilmeer/gists{/gist_id}", "starred_url": "https://api.github.com/users/usmandilmeer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/usmandilmeer/subscriptions", "organizations_url": "https://api.github.com/users/usmandilmeer/orgs", "repos_url": "https://api.github.com/users/usmandilmeer/repos", "events_url": "https://api.github.com/users/usmandilmeer/events{/privacy}", "received_events_url": "https://api.github.com/users/usmandilmeer/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4428/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4428/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5588
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5588/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5588/comments
https://api.github.com/repos/ollama/ollama/issues/5588/events
https://github.com/ollama/ollama/pull/5588
2,399,649,431
PR_kwDOJ0Z1Ps506G4S
5,588
Remove `GGML_CUDA_FORCE_MMQ=on` from build
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-10T03:08:09
2024-07-10T20:17:14
2024-07-10T20:17:13
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5588", "html_url": "https://github.com/ollama/ollama/pull/5588", "diff_url": "https://github.com/ollama/ollama/pull/5588.diff", "patch_url": "https://github.com/ollama/ollama/pull/5588.patch", "merged_at": "2024-07-10T20:17:13" }
Removing this build flag seems to cause mmq to be decided at runtime vs always forcing it on. This stops some cuda calls from failing on V100 and other CC 7.0 cards. This may also help with [RDNA3 cards](https://github.com/ggerganov/llama.cpp/blob/a59f8fdc85e1119d470d8766e29617962549d993/docs/build.md?plain=1#L175) Fixes https://github.com/ollama/ollama/issues/5571 Changes to mmq functionality related to this bug: https://github.com/ggerganov/llama.cpp/compare/7c26775...a8db2a9c#diff-b2fe862fcd5119199ae59ea13d1b6a46e0d23e41e727e39d90913f828a5ff66bR78
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5588/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5588/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1614
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1614/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1614/comments
https://api.github.com/repos/ollama/ollama/issues/1614/events
https://github.com/ollama/ollama/pull/1614
2,049,321,465
PR_kwDOJ0Z1Ps5iZ9OH
1,614
fix: set template without triple quotes
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-12-19T19:37:53
2024-01-09T17:36:26
2024-01-09T17:36:25
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1614", "html_url": "https://github.com/ollama/ollama/pull/1614", "diff_url": "https://github.com/ollama/ollama/pull/1614.diff", "patch_url": "https://github.com/ollama/ollama/pull/1614.patch", "merged_at": "2024-01-09T17:36:25" }
this changes updates `/set` to better handle multiline strings. `/set` now correctly sets template or system without using triple quotes ``` >>> /set template {{ .Prompt }} Set prompt template. ``` additionally, use a strings.Builder instead of concatenating string values for prompt building ``` >>> """hello ... world""" ``` ``` >>> """ ... hello ... world ... """ ``` ``` >>> /set system """ ... you are a llama ... """ ``` ``` >>> /set template """ ... {{.System}} ... User: {{.Prompt}} ... Assistant: {{.Response}} ... """ ``` resolves #1609 resolves #1607
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1614/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1614/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/473
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/473/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/473/comments
https://api.github.com/repos/ollama/ollama/issues/473/events
https://github.com/ollama/ollama/pull/473
1,882,917,488
PR_kwDOJ0Z1Ps5Zn-ay
473
create manifests directory
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-06T00:12:07
2023-09-06T00:37:42
2023-09-06T00:37:41
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/473", "html_url": "https://github.com/ollama/ollama/pull/473", "diff_url": "https://github.com/ollama/ollama/pull/473.diff", "patch_url": "https://github.com/ollama/ollama/pull/473.patch", "merged_at": "2023-09-06T00:37:41" }
`ollama list` on a brand new install will panic because the manifests directory doesn't exist
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/473/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/473/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1070
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1070/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1070/comments
https://api.github.com/repos/ollama/ollama/issues/1070/events
https://github.com/ollama/ollama/issues/1070
1,986,849,875
I_kwDOJ0Z1Ps52bOxT
1,070
API for models on `ollama.com`
{ "login": "danemadsen", "id": 11537699, "node_id": "MDQ6VXNlcjExNTM3Njk5", "avatar_url": "https://avatars.githubusercontent.com/u/11537699?v=4", "gravatar_id": "", "url": "https://api.github.com/users/danemadsen", "html_url": "https://github.com/danemadsen", "followers_url": "https://api.github.com/users/danemadsen/followers", "following_url": "https://api.github.com/users/danemadsen/following{/other_user}", "gists_url": "https://api.github.com/users/danemadsen/gists{/gist_id}", "starred_url": "https://api.github.com/users/danemadsen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/danemadsen/subscriptions", "organizations_url": "https://api.github.com/users/danemadsen/orgs", "repos_url": "https://api.github.com/users/danemadsen/repos", "events_url": "https://api.github.com/users/danemadsen/events{/privacy}", "received_events_url": "https://api.github.com/users/danemadsen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
closed
false
null
[]
null
12
2023-11-10T04:23:05
2024-11-21T17:25:46
2024-11-21T17:25:46
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi I was wondering if you could add a way to either search for, or get a list of models available to pull off ollama.ai. Currently the https://ollama.ai/library endpoint serves model information as HTML, it would be better if it was served as JSON so it could be better integrated into software
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1070/reactions", "total_count": 18, "+1": 18, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1070/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6120
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6120/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6120/comments
https://api.github.com/repos/ollama/ollama/issues/6120/events
https://github.com/ollama/ollama/issues/6120
2,442,618,478
I_kwDOJ0Z1Ps6Rl2Zu
6,120
Outputting the response leaves a bunch of control characters.
{ "login": "jjasghar", "id": 810824, "node_id": "MDQ6VXNlcjgxMDgyNA==", "avatar_url": "https://avatars.githubusercontent.com/u/810824?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jjasghar", "html_url": "https://github.com/jjasghar", "followers_url": "https://api.github.com/users/jjasghar/followers", "following_url": "https://api.github.com/users/jjasghar/following{/other_user}", "gists_url": "https://api.github.com/users/jjasghar/gists{/gist_id}", "starred_url": "https://api.github.com/users/jjasghar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jjasghar/subscriptions", "organizations_url": "https://api.github.com/users/jjasghar/orgs", "repos_url": "https://api.github.com/users/jjasghar/repos", "events_url": "https://api.github.com/users/jjasghar/events{/privacy}", "received_events_url": "https://api.github.com/users/jjasghar/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
9
2024-08-01T14:25:11
2024-11-22T16:04:55
2024-11-22T16:04:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I get a bunch of control characters. How can I get a clean output? _Originally posted by @pongnguy in https://github.com/ollama/ollama/issues/1710#issuecomment-2142294830_ It seems when you run something like: `ollama run granite-code:20b show me a python fuction that does fizzbuzz > out.txt` you end up with a bunch of: ``` ^[[?25l^[[?25hThis^[[?25l^[[?25h function^[[?25l^[[?25h takes^[[?2 5l^[[?25h a^[[?25l^[[?25h positive^[[?25l^[[?25h integer^[[?25l^[[ ?25h `^[[?25l^[[?25hn^[[?25l^[[?25h`^[[?25l^[[?25h as^[[?25l^[[?25 h input^[[?25l^[[?25h and^[[?25l^[[?25h prints^[[?25l^[[?25h the^[ [?25l^[[?25h numbers^[[?25l^[[?25h from^[[?25l^[[?25h ^[[?25l^[[?2 5h1^[[?25l^[[?25h to^[[?25l^[[?25h `^[[?25l^[[?25hn^[[?25l^[[?25h` ,^[[?25l^[[?25h replacing^[[?25l^[[?25h multip^[[?25l^[[?25hles^[[ ?25l^[[?25h of^[[?25l^[[?25h ^[[?25l^[[?25h3^[[?25l^[[?25h with^[[ ?25l^[[?25h "^[[?25l^[[?25hF^[[?25l^[[?25hizz^[[?25l^[[?25h",^[[?2 5l^[[?25h multip^[[?25l^[[?25hles^[[?25l^[[?25h of^[[?25l^[[?25h ^ [[?25l^[[?25h5^[[?25l^[[?25h with^[[?25l^[[?25h ``` Over and over. If you try to do the `--json` output, it's no better. It would be nice to have the simple text output with nothing else.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6120/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6120/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8127
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8127/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8127/comments
https://api.github.com/repos/ollama/ollama/issues/8127/events
https://github.com/ollama/ollama/pull/8127
2,743,777,670
PR_kwDOJ0Z1Ps6FboqQ
8,127
llm: loosen format check to default to no format
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-12-17T02:21:37
2024-12-17T04:25:55
2024-12-17T02:45:47
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8127", "html_url": "https://github.com/ollama/ollama/pull/8127", "diff_url": "https://github.com/ollama/ollama/pull/8127.diff", "patch_url": "https://github.com/ollama/ollama/pull/8127.patch", "merged_at": "2024-12-17T02:45:47" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8127/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8127/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1925
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1925/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1925/comments
https://api.github.com/repos/ollama/ollama/issues/1925/events
https://github.com/ollama/ollama/issues/1925
2,077,046,512
I_kwDOJ0Z1Ps57zTbw
1,925
Ollama is running in background in MacOS
{ "login": "MagzhanUnited", "id": 123943870, "node_id": "U_kgDOB2M7vg", "avatar_url": "https://avatars.githubusercontent.com/u/123943870?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MagzhanUnited", "html_url": "https://github.com/MagzhanUnited", "followers_url": "https://api.github.com/users/MagzhanUnited/followers", "following_url": "https://api.github.com/users/MagzhanUnited/following{/other_user}", "gists_url": "https://api.github.com/users/MagzhanUnited/gists{/gist_id}", "starred_url": "https://api.github.com/users/MagzhanUnited/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MagzhanUnited/subscriptions", "organizations_url": "https://api.github.com/users/MagzhanUnited/orgs", "repos_url": "https://api.github.com/users/MagzhanUnited/repos", "events_url": "https://api.github.com/users/MagzhanUnited/events{/privacy}", "received_events_url": "https://api.github.com/users/MagzhanUnited/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
11
2024-01-11T16:04:05
2025-01-29T08:51:22
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Even if I exit the ollama app I can see the ollama among run processes
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1925/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1925/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1806
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1806/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1806/comments
https://api.github.com/repos/ollama/ollama/issues/1806/events
https://github.com/ollama/ollama/issues/1806
2,067,309,672
I_kwDOJ0Z1Ps57OKRo
1,806
Add cli switch to show generation time and tokens/sec output time
{ "login": "thawkins", "id": 14537, "node_id": "MDQ6VXNlcjE0NTM3", "avatar_url": "https://avatars.githubusercontent.com/u/14537?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thawkins", "html_url": "https://github.com/thawkins", "followers_url": "https://api.github.com/users/thawkins/followers", "following_url": "https://api.github.com/users/thawkins/following{/other_user}", "gists_url": "https://api.github.com/users/thawkins/gists{/gist_id}", "starred_url": "https://api.github.com/users/thawkins/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thawkins/subscriptions", "organizations_url": "https://api.github.com/users/thawkins/orgs", "repos_url": "https://api.github.com/users/thawkins/repos", "events_url": "https://api.github.com/users/thawkins/events{/privacy}", "received_events_url": "https://api.github.com/users/thawkins/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-01-05T12:56:05
2024-01-08T18:56:01
2024-01-08T18:56:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Would it be possible to add a metrics switch to show net generation time and output time with tokens/seconds. This would make comparing the performance of LLMs easier.
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1806/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1806/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4451
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4451/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4451/comments
https://api.github.com/repos/ollama/ollama/issues/4451/events
https://github.com/ollama/ollama/pull/4451
2,297,709,787
PR_kwDOJ0Z1Ps5vhZ4t
4,451
add ability to create a client without env file
{ "login": "Tyrell04", "id": 43107913, "node_id": "MDQ6VXNlcjQzMTA3OTEz", "avatar_url": "https://avatars.githubusercontent.com/u/43107913?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Tyrell04", "html_url": "https://github.com/Tyrell04", "followers_url": "https://api.github.com/users/Tyrell04/followers", "following_url": "https://api.github.com/users/Tyrell04/following{/other_user}", "gists_url": "https://api.github.com/users/Tyrell04/gists{/gist_id}", "starred_url": "https://api.github.com/users/Tyrell04/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Tyrell04/subscriptions", "organizations_url": "https://api.github.com/users/Tyrell04/orgs", "repos_url": "https://api.github.com/users/Tyrell04/repos", "events_url": "https://api.github.com/users/Tyrell04/events{/privacy}", "received_events_url": "https://api.github.com/users/Tyrell04/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-05-15T11:57:58
2024-05-19T04:40:18
2024-05-18T03:08:30
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4451", "html_url": "https://github.com/ollama/ollama/pull/4451", "diff_url": "https://github.com/ollama/ollama/pull/4451.diff", "patch_url": "https://github.com/ollama/ollama/pull/4451.patch", "merged_at": null }
If you use Ollama as a Go package, you are forced to use an environment variable as a configuration to change the host. With the CreateClient function, you are free to use your own configuration passed into the Ollama client.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4451/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4451/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7324
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7324/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7324/comments
https://api.github.com/repos/ollama/ollama/issues/7324/events
https://github.com/ollama/ollama/pull/7324
2,606,172,201
PR_kwDOJ0Z1Ps5_f5eG
7,324
docs: use mingw32-make rather than MSYS Makefiles
{ "login": "shou692199", "id": 45505768, "node_id": "MDQ6VXNlcjQ1NTA1NzY4", "avatar_url": "https://avatars.githubusercontent.com/u/45505768?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shou692199", "html_url": "https://github.com/shou692199", "followers_url": "https://api.github.com/users/shou692199/followers", "following_url": "https://api.github.com/users/shou692199/following{/other_user}", "gists_url": "https://api.github.com/users/shou692199/gists{/gist_id}", "starred_url": "https://api.github.com/users/shou692199/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shou692199/subscriptions", "organizations_url": "https://api.github.com/users/shou692199/orgs", "repos_url": "https://api.github.com/users/shou692199/repos", "events_url": "https://api.github.com/users/shou692199/events{/privacy}", "received_events_url": "https://api.github.com/users/shou692199/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-22T18:26:14
2024-10-31T13:39:53
2024-10-31T13:39:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7324", "html_url": "https://github.com/ollama/ollama/pull/7324", "diff_url": "https://github.com/ollama/ollama/pull/7324.diff", "patch_url": "https://github.com/ollama/ollama/pull/7324.patch", "merged_at": null }
Since I'm trying to port ggml-sycl to llama package, MSYS Makefiles doesn't support icx compiler but MinGW Makefiles does, and it also works well with cpu, rocm and cuda. I think it's an ideal choice.
{ "login": "shou692199", "id": 45505768, "node_id": "MDQ6VXNlcjQ1NTA1NzY4", "avatar_url": "https://avatars.githubusercontent.com/u/45505768?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shou692199", "html_url": "https://github.com/shou692199", "followers_url": "https://api.github.com/users/shou692199/followers", "following_url": "https://api.github.com/users/shou692199/following{/other_user}", "gists_url": "https://api.github.com/users/shou692199/gists{/gist_id}", "starred_url": "https://api.github.com/users/shou692199/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shou692199/subscriptions", "organizations_url": "https://api.github.com/users/shou692199/orgs", "repos_url": "https://api.github.com/users/shou692199/repos", "events_url": "https://api.github.com/users/shou692199/events{/privacy}", "received_events_url": "https://api.github.com/users/shou692199/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7324/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7324/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2292
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2292/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2292/comments
https://api.github.com/repos/ollama/ollama/issues/2292/events
https://github.com/ollama/ollama/issues/2292
2,110,861,549
I_kwDOJ0Z1Ps590TDt
2,292
Manifest file?
{ "login": "EquaTechnologies", "id": 107852330, "node_id": "U_kgDOBm2yKg", "avatar_url": "https://avatars.githubusercontent.com/u/107852330?v=4", "gravatar_id": "", "url": "https://api.github.com/users/EquaTechnologies", "html_url": "https://github.com/EquaTechnologies", "followers_url": "https://api.github.com/users/EquaTechnologies/followers", "following_url": "https://api.github.com/users/EquaTechnologies/following{/other_user}", "gists_url": "https://api.github.com/users/EquaTechnologies/gists{/gist_id}", "starred_url": "https://api.github.com/users/EquaTechnologies/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/EquaTechnologies/subscriptions", "organizations_url": "https://api.github.com/users/EquaTechnologies/orgs", "repos_url": "https://api.github.com/users/EquaTechnologies/repos", "events_url": "https://api.github.com/users/EquaTechnologies/events{/privacy}", "received_events_url": "https://api.github.com/users/EquaTechnologies/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-31T19:57:51
2024-02-20T04:05:47
2024-02-20T04:05:47
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
qua@equa-Swift-SF314-54:~$ ollama run orca pulling manifest Error: pull model manifest: file does not exist equa@equa-Swift-SF314-54:~$
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2292/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2292/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6240
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6240/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6240/comments
https://api.github.com/repos/ollama/ollama/issues/6240/events
https://github.com/ollama/ollama/issues/6240
2,454,100,567
I_kwDOJ0Z1Ps6SRppX
6,240
Not executed in gpu amd rx 6750 GRE
{ "login": "21307369", "id": 47931342, "node_id": "MDQ6VXNlcjQ3OTMxMzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47931342?v=4", "gravatar_id": "", "url": "https://api.github.com/users/21307369", "html_url": "https://github.com/21307369", "followers_url": "https://api.github.com/users/21307369/followers", "following_url": "https://api.github.com/users/21307369/following{/other_user}", "gists_url": "https://api.github.com/users/21307369/gists{/gist_id}", "starred_url": "https://api.github.com/users/21307369/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/21307369/subscriptions", "organizations_url": "https://api.github.com/users/21307369/orgs", "repos_url": "https://api.github.com/users/21307369/repos", "events_url": "https://api.github.com/users/21307369/events{/privacy}", "received_events_url": "https://api.github.com/users/21307369/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
3
2024-08-07T18:39:39
2024-09-03T19:10:36
2024-08-09T19:25:52
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
<img width="898" alt="image" src="https://github.com/user-attachments/assets/87a706b4-7749-443f-86ea-86a3c7de1cc1"> 2024/08/08 14:17:22 routes.go:1108: INFO server config env="map[OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/Users/lsmir2/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[* http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_RUNNERS_DIR: OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR:]" time=2024-08-08T14:17:22.106+08:00 level=INFO source=images.go:781 msg="total blobs: 5" time=2024-08-08T14:17:22.106+08:00 level=INFO source=images.go:788 msg="total unused blobs removed: 0" time=2024-08-08T14:17:22.107+08:00 level=INFO source=routes.go:1155 msg="Listening on 127.0.0.1:11434 (version 0.3.4)" time=2024-08-08T14:17:22.113+08:00 level=WARN source=assets.go:100 msg="unable to cleanup stale tmpdir" path=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama1166271814 error="remove /var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama1166271814: directory not empty" time=2024-08-08T14:17:22.114+08:00 level=WARN source=assets.go:100 msg="unable to cleanup stale tmpdir" path=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama2116948276 error="remove /var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama2116948276: directory not empty" time=2024-08-08T14:17:22.114+08:00 level=WARN source=assets.go:100 msg="unable to cleanup stale tmpdir" path=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama3670480486 error="remove /var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama3670480486: directory not empty" time=2024-08-08T14:17:22.115+08:00 level=INFO source=payload.go:30 msg="extracting embedded files" dir=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama4115735691/runners time=2024-08-08T14:17:22.146+08:00 level=INFO source=payload.go:44 msg="Dynamic LLM libraries [cpu cpu_avx cpu_avx2]" time=2024-08-08T14:17:22.146+08:00 level=INFO source=types.go:105 msg="inference compute" id="" library=cpu compute="" driver=0.0 name="" total="32.0 GiB" available="16.1 GiB" time=2024-08-08T14:17:24.808+08:00 level=INFO source=memory.go:309 msg="offload to cpu" layers.requested=-1 layers.model=41 layers.offload=0 layers.split="" memory.available="[16.1 GiB]" memory.required.full="5.7 GiB" memory.required.partial="0 B" memory.required.kv="320.0 MiB" memory.required.allocations="[5.7 GiB]" memory.weights.total="4.6 GiB" memory.weights.repeating="4.1 GiB" memory.weights.nonrepeating="485.6 MiB" memory.graph.full="561.0 MiB" memory.graph.partial="789.6 MiB" time=2024-08-08T14:17:24.809+08:00 level=INFO source=server.go:392 msg="starting llama server" cmd="/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama4115735691/runners/cpu_avx2/ollama_llama_server --model /Users/lsmir2/.ollama/models/blobs/sha256-816441b33390807d429fbdb1de7e33bb4d569ac68e2203bdbca5d8d79b5c7266 --ctx-size 8192 --batch-size 512 --embedding --log-disable --no-mmap --parallel 4 --port 57883" time=2024-08-08T14:17:24.816+08:00 level=INFO source=sched.go:445 msg="loaded runners" count=1 time=2024-08-08T14:17:24.816+08:00 level=INFO source=server.go:592 msg="waiting for llama runner to start responding" time=2024-08-08T14:17:24.817+08:00 level=INFO source=server.go:626 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=3535 commit="1e6f6554" tid="0x107f92600" timestamp=1723097844 INFO [main] system info | n_threads=6 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="0x107f92600" timestamp=1723097844 total_threads=6 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="6" port="57883" tid="0x107f92600" timestamp=1723097844 llama_model_loader: loaded meta data with 23 key-value pairs and 283 tensors from /Users/lsmir2/.ollama/models/blobs/sha256-816441b33390807d429fbdb1de7e33bb4d569ac68e2203bdbca5d8d79b5c7266 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = chatglm llama_model_loader: - kv 1: general.name str = codegeex4-all-9b llama_model_loader: - kv 2: chatglm.context_length u32 = 131072 llama_model_loader: - kv 3: chatglm.embedding_length u32 = 4096 llama_model_loader: - kv 4: chatglm.feed_forward_length u32 = 13696 llama_model_loader: - kv 5: chatglm.block_count u32 = 40 llama_model_loader: - kv 6: chatglm.attention.head_count u32 = 32 llama_model_loader: - kv 7: chatglm.attention.head_count_kv u32 = 2 llama_model_loader: - kv 8: chatglm.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 9: general.file_type u32 = 2 llama_model_loader: - kv 10: chatglm.rope.dimension_count u32 = 64 llama_model_loader: - kv 11: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 12: chatglm.rope.freq_base f32 = 5000000.000000 llama_model_loader: - kv 13: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 14: tokenizer.ggml.pre str = chatglm-bpe llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,151552] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 16: tokenizer.ggml.token_type arr[i32,151552] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 17: tokenizer.ggml.merges arr[str,151073] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 18: tokenizer.ggml.padding_token_id u32 = 151329 llama_model_loader: - kv 19: tokenizer.ggml.eos_token_id u32 = 151329 llama_model_loader: - kv 20: tokenizer.ggml.eot_token_id u32 = 151336 llama_model_loader: - kv 21: tokenizer.ggml.unknown_token_id u32 = 151329 llama_model_loader: - kv 22: general.quantization_version u32 = 2 llama_model_loader: - type f32: 121 tensors llama_model_loader: - type q4_0: 161 tensors llama_model_loader: - type q6_K: 1 tensors time=2024-08-08T14:17:25.069+08:00 level=INFO source=server.go:626 msg="waiting for server to become available" status="llm server loading model" llm_load_vocab: special tokens cache size = 223 llm_load_vocab: token to piece cache size = 0.9732 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = chatglm llm_load_print_meta: vocab type = BPE llm_load_print_meta: n_vocab = 151552 llm_load_print_meta: n_merges = 151073 llm_load_print_meta: vocab_only = 0 llm_load_print_meta: n_ctx_train = 131072 llm_load_print_meta: n_embd = 4096 llm_load_print_meta: n_layer = 40 llm_load_print_meta: n_head = 32 llm_load_print_meta: n_head_kv = 2 llm_load_print_meta: n_rot = 64 llm_load_print_meta: n_swa = 0 llm_load_print_meta: n_embd_head_k = 128 llm_load_print_meta: n_embd_head_v = 128 llm_load_print_meta: n_gqa = 16 llm_load_print_meta: n_embd_k_gqa = 256 llm_load_print_meta: n_embd_v_gqa = 256 llm_load_print_meta: f_norm_eps = 0.0e+00 llm_load_print_meta: f_norm_rms_eps = 1.0e-05 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 13696 llm_load_print_meta: n_expert = 0 llm_load_print_meta: n_expert_used = 0 llm_load_print_meta: causal attn = 1 llm_load_print_meta: pooling type = 0 llm_load_print_meta: rope type = 0 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 5000000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_ctx_orig_yarn = 131072 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: model type = 9B llm_load_print_meta: model ftype = Q4_0 llm_load_print_meta: model params = 9.40 B llm_load_print_meta: model size = 5.08 GiB (4.64 BPW) llm_load_print_meta: general.name = codegeex4-all-9b llm_load_print_meta: EOS token = 151329 '<|endoftext|>' llm_load_print_meta: UNK token = 151329 '<|endoftext|>' llm_load_print_meta: PAD token = 151329 '<|endoftext|>' llm_load_print_meta: LF token = 128 'Ä' llm_load_print_meta: EOT token = 151336 '<|user|>' llm_load_print_meta: max token length = 1024 llm_load_tensors: ggml ctx size = 0.14 MiB llm_load_tensors: CPU buffer size = 5196.84 MiB llama_new_context_with_model: n_ctx = 8192 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 5000000.0 llama_new_context_with_model: freq_scale = 1 llama_kv_cache_init: CPU KV buffer size = 320.00 MiB llama_new_context_with_model: KV self size = 320.00 MiB, K (f16): 160.00 MiB, V (f16): 160.00 MiB llama_new_context_with_model: CPU output buffer size = 2.38 MiB llama_new_context_with_model: CPU compute buffer size = 561.01 MiB llama_new_context_with_model: graph nodes = 1606 llama_new_context_with_model: graph splits = 1 INFO [main] model loaded | tid="0x107f92600" timestamp=1723097848 time=2024-08-08T14:17:28.352+08:00 level=INFO source=server.go:631 msg="llama runner started in 3.54 seconds" [GIN] 2024/08/08 - 14:17:33 | 200 | 8.515598533s | 127.0.0.1 | POST "/v1/chat/completions"
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6240/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6240/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5511
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5511/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5511/comments
https://api.github.com/repos/ollama/ollama/issues/5511/events
https://github.com/ollama/ollama/pull/5511
2,393,285,539
PR_kwDOJ0Z1Ps50kwRU
5,511
llm: fix missing dylibs by restoring old build behavior on Linux and macOS
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-06T01:00:34
2024-07-08T16:55:51
2024-07-06T01:48:32
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5511", "html_url": "https://github.com/ollama/ollama/pull/5511", "diff_url": "https://github.com/ollama/ollama/pull/5511.diff", "patch_url": "https://github.com/ollama/ollama/pull/5511.patch", "merged_at": "2024-07-06T01:48:32" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5511/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5511/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7660
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7660/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7660/comments
https://api.github.com/repos/ollama/ollama/issues/7660/events
https://github.com/ollama/ollama/issues/7660
2,657,322,190
I_kwDOJ0Z1Ps6eY4TO
7,660
on ollama model specifics web sites , add comaptible copypaste <name>:<tag> text
{ "login": "fxmbsw7", "id": 39368685, "node_id": "MDQ6VXNlcjM5MzY4Njg1", "avatar_url": "https://avatars.githubusercontent.com/u/39368685?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fxmbsw7", "html_url": "https://github.com/fxmbsw7", "followers_url": "https://api.github.com/users/fxmbsw7/followers", "following_url": "https://api.github.com/users/fxmbsw7/following{/other_user}", "gists_url": "https://api.github.com/users/fxmbsw7/gists{/gist_id}", "starred_url": "https://api.github.com/users/fxmbsw7/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fxmbsw7/subscriptions", "organizations_url": "https://api.github.com/users/fxmbsw7/orgs", "repos_url": "https://api.github.com/users/fxmbsw7/repos", "events_url": "https://api.github.com/users/fxmbsw7/events{/privacy}", "received_events_url": "https://api.github.com/users/fxmbsw7/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
closed
false
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
9
2024-11-14T02:16:13
2024-12-01T07:31:54
2024-11-15T02:24:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
eg , user visits page of a model chooses select list , and selects some model version there on that version of model page i want say , for qwen2.5-coder.on its 14b-instruct-q4_K_M page simply qwen2.5-coder:14b-instruct-q4_K_M name:tag like hf s use that button with ollama copypaste run cmd in oh and yea , peeferably also a small ' copy me ' button left on it greets
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7660/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7660/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/630
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/630/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/630/comments
https://api.github.com/repos/ollama/ollama/issues/630/events
https://github.com/ollama/ollama/issues/630
1,916,976,222
I_kwDOJ0Z1Ps5yQrxe
630
Error: failed to start a llama runner
{ "login": "azhang", "id": 940275, "node_id": "MDQ6VXNlcjk0MDI3NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/940275?v=4", "gravatar_id": "", "url": "https://api.github.com/users/azhang", "html_url": "https://github.com/azhang", "followers_url": "https://api.github.com/users/azhang/followers", "following_url": "https://api.github.com/users/azhang/following{/other_user}", "gists_url": "https://api.github.com/users/azhang/gists{/gist_id}", "starred_url": "https://api.github.com/users/azhang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/azhang/subscriptions", "organizations_url": "https://api.github.com/users/azhang/orgs", "repos_url": "https://api.github.com/users/azhang/repos", "events_url": "https://api.github.com/users/azhang/events{/privacy}", "received_events_url": "https://api.github.com/users/azhang/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
28
2023-09-28T08:25:32
2024-07-12T19:18:02
2023-10-12T15:16:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When I run ollama run mistral it downloads properly but then fails to run it, with the following error: Error: failed to start a llama runner I'm running this on my intel mbp with 64g ram
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/630/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/630/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3661
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3661/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3661/comments
https://api.github.com/repos/ollama/ollama/issues/3661/events
https://github.com/ollama/ollama/pull/3661
2,244,821,343
PR_kwDOJ0Z1Ps5svQOm
3,661
better checking for OLLAMA_HOST variable
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-15T23:39:21
2024-04-29T23:14:08
2024-04-29T23:14:07
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3661", "html_url": "https://github.com/ollama/ollama/pull/3661", "diff_url": "https://github.com/ollama/ollama/pull/3661.diff", "patch_url": "https://github.com/ollama/ollama/pull/3661.patch", "merged_at": "2024-04-29T23:14:07" }
This change adds better validation to the `OLLAMA_HOST` variable when used with `ollama serve`. It should work with both IPv4 and IPv6, and includes unit tests.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3661/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3661/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1109
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1109/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1109/comments
https://api.github.com/repos/ollama/ollama/issues/1109/events
https://github.com/ollama/ollama/issues/1109
1,990,754,546
I_kwDOJ0Z1Ps52qIDy
1,109
Exploring Multimodal LLMs: Incorporating Image as Input in ModelFile.
{ "login": "saifullah-N", "id": 79261602, "node_id": "MDQ6VXNlcjc5MjYxNjAy", "avatar_url": "https://avatars.githubusercontent.com/u/79261602?v=4", "gravatar_id": "", "url": "https://api.github.com/users/saifullah-N", "html_url": "https://github.com/saifullah-N", "followers_url": "https://api.github.com/users/saifullah-N/followers", "following_url": "https://api.github.com/users/saifullah-N/following{/other_user}", "gists_url": "https://api.github.com/users/saifullah-N/gists{/gist_id}", "starred_url": "https://api.github.com/users/saifullah-N/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/saifullah-N/subscriptions", "organizations_url": "https://api.github.com/users/saifullah-N/orgs", "repos_url": "https://api.github.com/users/saifullah-N/repos", "events_url": "https://api.github.com/users/saifullah-N/events{/privacy}", "received_events_url": "https://api.github.com/users/saifullah-N/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-11-13T14:18:07
2023-11-13T17:42:57
2023-11-13T17:42:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, I'd like to experiment with multimodal language models that can handle both images and text as input. Is there a way to input an image. My ModelFile looks like this ``` FROM ./my_model.gguf PARAMERTER.. SYSTEM You are An Artist describe the image ``` I'm trying to run this on macos. M1 chip
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1109/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1109/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3728
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3728/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3728/comments
https://api.github.com/repos/ollama/ollama/issues/3728/events
https://github.com/ollama/ollama/issues/3728
2,249,868,609
I_kwDOJ0Z1Ps6GGkVB
3,728
ollama prune
{ "login": "Tobiadefami", "id": 48369656, "node_id": "MDQ6VXNlcjQ4MzY5NjU2", "avatar_url": "https://avatars.githubusercontent.com/u/48369656?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Tobiadefami", "html_url": "https://github.com/Tobiadefami", "followers_url": "https://api.github.com/users/Tobiadefami/followers", "following_url": "https://api.github.com/users/Tobiadefami/following{/other_user}", "gists_url": "https://api.github.com/users/Tobiadefami/gists{/gist_id}", "starred_url": "https://api.github.com/users/Tobiadefami/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Tobiadefami/subscriptions", "organizations_url": "https://api.github.com/users/Tobiadefami/orgs", "repos_url": "https://api.github.com/users/Tobiadefami/repos", "events_url": "https://api.github.com/users/Tobiadefami/events{/privacy}", "received_events_url": "https://api.github.com/users/Tobiadefami/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-04-18T06:35:53
2024-04-18T08:19:21
2024-04-18T08:19:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I think ollama needs a prune command, something similar to `docker system prune`. Just to keep the environment tidy
{ "login": "Tobiadefami", "id": 48369656, "node_id": "MDQ6VXNlcjQ4MzY5NjU2", "avatar_url": "https://avatars.githubusercontent.com/u/48369656?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Tobiadefami", "html_url": "https://github.com/Tobiadefami", "followers_url": "https://api.github.com/users/Tobiadefami/followers", "following_url": "https://api.github.com/users/Tobiadefami/following{/other_user}", "gists_url": "https://api.github.com/users/Tobiadefami/gists{/gist_id}", "starred_url": "https://api.github.com/users/Tobiadefami/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Tobiadefami/subscriptions", "organizations_url": "https://api.github.com/users/Tobiadefami/orgs", "repos_url": "https://api.github.com/users/Tobiadefami/repos", "events_url": "https://api.github.com/users/Tobiadefami/events{/privacy}", "received_events_url": "https://api.github.com/users/Tobiadefami/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3728/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3728/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3660
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3660/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3660/comments
https://api.github.com/repos/ollama/ollama/issues/3660/events
https://github.com/ollama/ollama/issues/3660
2,244,762,677
I_kwDOJ0Z1Ps6FzFw1
3,660
Exception using command-r-plus:104b-q4_0
{ "login": "mehdiataei", "id": 29686241, "node_id": "MDQ6VXNlcjI5Njg2MjQx", "avatar_url": "https://avatars.githubusercontent.com/u/29686241?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mehdiataei", "html_url": "https://github.com/mehdiataei", "followers_url": "https://api.github.com/users/mehdiataei/followers", "following_url": "https://api.github.com/users/mehdiataei/following{/other_user}", "gists_url": "https://api.github.com/users/mehdiataei/gists{/gist_id}", "starred_url": "https://api.github.com/users/mehdiataei/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mehdiataei/subscriptions", "organizations_url": "https://api.github.com/users/mehdiataei/orgs", "repos_url": "https://api.github.com/users/mehdiataei/repos", "events_url": "https://api.github.com/users/mehdiataei/events{/privacy}", "received_events_url": "https://api.github.com/users/mehdiataei/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
7
2024-04-15T22:49:24
2024-04-17T00:41:10
2024-04-17T00:41:10
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Ollama: exception done_getting_tensors: wrong number of tensors; expected 642, got 514 ### What did you expect to see? _No response_ ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS _No response_ ### Architecture _No response_ ### Platform _No response_ ### Ollama version _No response_ ### GPU _No response_ ### GPU info _No response_ ### CPU _No response_ ### Other software _No response_
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3660/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3660/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2792
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2792/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2792/comments
https://api.github.com/repos/ollama/ollama/issues/2792/events
https://github.com/ollama/ollama/issues/2792
2,157,751,613
I_kwDOJ0Z1Ps6AnK09
2,792
Subsequent generation requests hang after successful generation request with `num_predict: 0`
{ "login": "stanier", "id": 8770519, "node_id": "MDQ6VXNlcjg3NzA1MTk=", "avatar_url": "https://avatars.githubusercontent.com/u/8770519?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stanier", "html_url": "https://github.com/stanier", "followers_url": "https://api.github.com/users/stanier/followers", "following_url": "https://api.github.com/users/stanier/following{/other_user}", "gists_url": "https://api.github.com/users/stanier/gists{/gist_id}", "starred_url": "https://api.github.com/users/stanier/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stanier/subscriptions", "organizations_url": "https://api.github.com/users/stanier/orgs", "repos_url": "https://api.github.com/users/stanier/repos", "events_url": "https://api.github.com/users/stanier/events{/privacy}", "received_events_url": "https://api.github.com/users/stanier/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-27T22:54:24
2024-05-10T01:13:35
2024-05-10T01:13:35
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
If you submit a generation request with `num_predict: 0`, the request will be handled successfully but all subsequent generation requests will hang indefinitely regardless of their `num_predict` values. Below is an example of how to reproduce this behavior: ``` curl http://localhost:11434/api/generate -d '{ "model": "vicuna:13b-16k", "template":"Hello world!", "stream":false, "num_predict":0 }' ``` Any requests following this will not return until the daemon has been restarted. Also to note is that the daemon might be in an unresponsive state following this request, I've had to `kill` it each time it's happened to me so far, but haven't tested this against typical behavior for signals sent mid-generation or otherwise. I think it also affects `raw` mode, and neither `num_keep` or `num_ctx` seem to be related.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2792/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2792/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1849
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1849/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1849/comments
https://api.github.com/repos/ollama/ollama/issues/1849/events
https://github.com/ollama/ollama/pull/1849
2,069,374,142
PR_kwDOJ0Z1Ps5jatea
1,849
Accomodate split cuda lib dir
{ "login": "mraiser", "id": 1268027, "node_id": "MDQ6VXNlcjEyNjgwMjc=", "avatar_url": "https://avatars.githubusercontent.com/u/1268027?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mraiser", "html_url": "https://github.com/mraiser", "followers_url": "https://api.github.com/users/mraiser/followers", "following_url": "https://api.github.com/users/mraiser/following{/other_user}", "gists_url": "https://api.github.com/users/mraiser/gists{/gist_id}", "starred_url": "https://api.github.com/users/mraiser/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mraiser/subscriptions", "organizations_url": "https://api.github.com/users/mraiser/orgs", "repos_url": "https://api.github.com/users/mraiser/repos", "events_url": "https://api.github.com/users/mraiser/events{/privacy}", "received_events_url": "https://api.github.com/users/mraiser/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
9
2024-01-08T00:24:45
2024-02-06T00:01:17
2024-02-06T00:01:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1849", "html_url": "https://github.com/ollama/ollama/pull/1849", "diff_url": "https://github.com/ollama/ollama/pull/1849.diff", "patch_url": "https://github.com/ollama/ollama/pull/1849.patch", "merged_at": "2024-02-06T00:01:17" }
Makes it a little easier to compile when cuda lib dir is split up as in nixos.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1849/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1849/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6786
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6786/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6786/comments
https://api.github.com/repos/ollama/ollama/issues/6786/events
https://github.com/ollama/ollama/issues/6786
2,523,912,567
I_kwDOJ0Z1Ps6Wb9l3
6,786
Isn't it time to move onto Omni models?
{ "login": "Meshwa428", "id": 135232056, "node_id": "U_kgDOCA96OA", "avatar_url": "https://avatars.githubusercontent.com/u/135232056?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Meshwa428", "html_url": "https://github.com/Meshwa428", "followers_url": "https://api.github.com/users/Meshwa428/followers", "following_url": "https://api.github.com/users/Meshwa428/following{/other_user}", "gists_url": "https://api.github.com/users/Meshwa428/gists{/gist_id}", "starred_url": "https://api.github.com/users/Meshwa428/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Meshwa428/subscriptions", "organizations_url": "https://api.github.com/users/Meshwa428/orgs", "repos_url": "https://api.github.com/users/Meshwa428/repos", "events_url": "https://api.github.com/users/Meshwa428/events{/privacy}", "received_events_url": "https://api.github.com/users/Meshwa428/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-09-13T05:34:46
2024-10-23T17:17:40
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
There is a model that I found today called *LLaMa 3.1 8b Omni* and it is a speech to speech model with very low latency ensuring the best experience for local models. But as for Ollama it doesn't support such models, although VLMs are there but having these Omni models on your local device is just 🤌🏻 too Good to be true. HF reference for llama 3.1 8b Omni: https://huggingface.co/ICTNLP/Llama-3.1-8B-Omni Looking forward to @ollama team's implementation Thanks🙏🏻
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6786/reactions", "total_count": 9, "+1": 9, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6786/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5244
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5244/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5244/comments
https://api.github.com/repos/ollama/ollama/issues/5244/events
https://github.com/ollama/ollama/pull/5244
2,368,887,556
PR_kwDOJ0Z1Ps5zS5Hb
5,244
llm: suppress large allocations for GGUF arrays
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-06-23T21:01:06
2024-06-23T22:32:04
2024-06-23T22:32:04
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
true
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5244", "html_url": "https://github.com/ollama/ollama/pull/5244", "diff_url": "https://github.com/ollama/ollama/pull/5244.diff", "patch_url": "https://github.com/ollama/ollama/pull/5244.patch", "merged_at": null }
This introduces a little array type for holding GGUF arrays that prevents the array from growing too large. It preserves the total size of the array, but limits the number of elements that are actually allocated. GGUF arrays that are extremely large, such as tokens, etc, are generally uninteresting to users, and are not worth the memory overhead, and the time spent allocating and freeing them. They are necessary for inference, but not for inspection. The size of these arrays is, however, important in Ollama, so it is preserved in a separate field on array.
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5244/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5244/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/836
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/836/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/836/comments
https://api.github.com/repos/ollama/ollama/issues/836/events
https://github.com/ollama/ollama/issues/836
1,949,521,611
I_kwDOJ0Z1Ps50M1bL
836
Support setting parameters without a modelfile
{ "login": "chigkim", "id": 22120994, "node_id": "MDQ6VXNlcjIyMTIwOTk0", "avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chigkim", "html_url": "https://github.com/chigkim", "followers_url": "https://api.github.com/users/chigkim/followers", "following_url": "https://api.github.com/users/chigkim/following{/other_user}", "gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}", "starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chigkim/subscriptions", "organizations_url": "https://api.github.com/users/chigkim/orgs", "repos_url": "https://api.github.com/users/chigkim/repos", "events_url": "https://api.github.com/users/chigkim/events{/privacy}", "received_events_url": "https://api.github.com/users/chigkim/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2023-10-18T11:30:51
2024-05-05T23:14:26
2024-01-03T03:38:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I know you can set temperature using modelfile or api, but it would be much more convenient to be able to do it in cli. For example: `ollama run llama-2 -temperture 0.1` when loading and/or `/set temperature 0.1` in runtime Thanks for your consideration!
{ "login": "chigkim", "id": 22120994, "node_id": "MDQ6VXNlcjIyMTIwOTk0", "avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chigkim", "html_url": "https://github.com/chigkim", "followers_url": "https://api.github.com/users/chigkim/followers", "following_url": "https://api.github.com/users/chigkim/following{/other_user}", "gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}", "starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chigkim/subscriptions", "organizations_url": "https://api.github.com/users/chigkim/orgs", "repos_url": "https://api.github.com/users/chigkim/repos", "events_url": "https://api.github.com/users/chigkim/events{/privacy}", "received_events_url": "https://api.github.com/users/chigkim/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/836/reactions", "total_count": 5, "+1": 5, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/836/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4747
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4747/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4747/comments
https://api.github.com/repos/ollama/ollama/issues/4747/events
https://github.com/ollama/ollama/issues/4747
2,327,194,470
I_kwDOJ0Z1Ps6Ktitm
4,747
Running multiple models simultaneously, always using one card
{ "login": "leoHostProject", "id": 87935281, "node_id": "MDQ6VXNlcjg3OTM1Mjgx", "avatar_url": "https://avatars.githubusercontent.com/u/87935281?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leoHostProject", "html_url": "https://github.com/leoHostProject", "followers_url": "https://api.github.com/users/leoHostProject/followers", "following_url": "https://api.github.com/users/leoHostProject/following{/other_user}", "gists_url": "https://api.github.com/users/leoHostProject/gists{/gist_id}", "starred_url": "https://api.github.com/users/leoHostProject/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/leoHostProject/subscriptions", "organizations_url": "https://api.github.com/users/leoHostProject/orgs", "repos_url": "https://api.github.com/users/leoHostProject/repos", "events_url": "https://api.github.com/users/leoHostProject/events{/privacy}", "received_events_url": "https://api.github.com/users/leoHostProject/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-05-31T07:58:11
2024-05-31T18:58:02
2024-05-31T18:57:10
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Running multiple models simultaneously, always using one card,but i have 4 cards and download 4model When multiple users are using it at the same time, always clear the first card and then load other models instead of using my other idle cards ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.33
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4747/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4747/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4583
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4583/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4583/comments
https://api.github.com/repos/ollama/ollama/issues/4583/events
https://github.com/ollama/ollama/pull/4583
2,311,914,296
PR_kwDOJ0Z1Ps5wRyOi
4,583
Add new community integration (TypingMind)
{ "login": "trungdq88", "id": 4214509, "node_id": "MDQ6VXNlcjQyMTQ1MDk=", "avatar_url": "https://avatars.githubusercontent.com/u/4214509?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trungdq88", "html_url": "https://github.com/trungdq88", "followers_url": "https://api.github.com/users/trungdq88/followers", "following_url": "https://api.github.com/users/trungdq88/following{/other_user}", "gists_url": "https://api.github.com/users/trungdq88/gists{/gist_id}", "starred_url": "https://api.github.com/users/trungdq88/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/trungdq88/subscriptions", "organizations_url": "https://api.github.com/users/trungdq88/orgs", "repos_url": "https://api.github.com/users/trungdq88/repos", "events_url": "https://api.github.com/users/trungdq88/events{/privacy}", "received_events_url": "https://api.github.com/users/trungdq88/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-05-23T04:31:39
2024-06-10T09:27:58
2024-06-10T09:27:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4583", "html_url": "https://github.com/ollama/ollama/pull/4583", "diff_url": "https://github.com/ollama/ollama/pull/4583.diff", "patch_url": "https://github.com/ollama/ollama/pull/4583.patch", "merged_at": null }
Cheers!
{ "login": "trungdq88", "id": 4214509, "node_id": "MDQ6VXNlcjQyMTQ1MDk=", "avatar_url": "https://avatars.githubusercontent.com/u/4214509?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trungdq88", "html_url": "https://github.com/trungdq88", "followers_url": "https://api.github.com/users/trungdq88/followers", "following_url": "https://api.github.com/users/trungdq88/following{/other_user}", "gists_url": "https://api.github.com/users/trungdq88/gists{/gist_id}", "starred_url": "https://api.github.com/users/trungdq88/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/trungdq88/subscriptions", "organizations_url": "https://api.github.com/users/trungdq88/orgs", "repos_url": "https://api.github.com/users/trungdq88/repos", "events_url": "https://api.github.com/users/trungdq88/events{/privacy}", "received_events_url": "https://api.github.com/users/trungdq88/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4583/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4583/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5688
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5688/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5688/comments
https://api.github.com/repos/ollama/ollama/issues/5688/events
https://github.com/ollama/ollama/pull/5688
2,407,318,220
PR_kwDOJ0Z1Ps51T_VY
5,688
Serve static files
{ "login": "1feralcat", "id": 51179976, "node_id": "MDQ6VXNlcjUxMTc5OTc2", "avatar_url": "https://avatars.githubusercontent.com/u/51179976?v=4", "gravatar_id": "", "url": "https://api.github.com/users/1feralcat", "html_url": "https://github.com/1feralcat", "followers_url": "https://api.github.com/users/1feralcat/followers", "following_url": "https://api.github.com/users/1feralcat/following{/other_user}", "gists_url": "https://api.github.com/users/1feralcat/gists{/gist_id}", "starred_url": "https://api.github.com/users/1feralcat/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/1feralcat/subscriptions", "organizations_url": "https://api.github.com/users/1feralcat/orgs", "repos_url": "https://api.github.com/users/1feralcat/repos", "events_url": "https://api.github.com/users/1feralcat/events{/privacy}", "received_events_url": "https://api.github.com/users/1feralcat/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-14T08:06:08
2024-08-14T16:46:19
2024-08-14T16:46:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5688", "html_url": "https://github.com/ollama/ollama/pull/5688", "diff_url": "https://github.com/ollama/ollama/pull/5688.diff", "patch_url": "https://github.com/ollama/ollama/pull/5688.patch", "merged_at": null }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5688/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5688/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5707
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5707/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5707/comments
https://api.github.com/repos/ollama/ollama/issues/5707/events
https://github.com/ollama/ollama/pull/5707
2,409,438,713
PR_kwDOJ0Z1Ps51bLDN
5,707
Image Embeddings
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-15T19:13:38
2024-11-21T10:11:28
2024-11-21T10:11:28
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
true
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5707", "html_url": "https://github.com/ollama/ollama/pull/5707", "diff_url": "https://github.com/ollama/ollama/pull/5707.diff", "patch_url": "https://github.com/ollama/ollama/pull/5707.patch", "merged_at": null }
rough work
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5707/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5707/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6117
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6117/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6117/comments
https://api.github.com/repos/ollama/ollama/issues/6117/events
https://github.com/ollama/ollama/issues/6117
2,442,322,095
I_kwDOJ0Z1Ps6RkuCv
6,117
Add Gemma 2 2b base/ text/ pre-trained model to registry
{ "login": "nviraj", "id": 8409854, "node_id": "MDQ6VXNlcjg0MDk4NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8409854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nviraj", "html_url": "https://github.com/nviraj", "followers_url": "https://api.github.com/users/nviraj/followers", "following_url": "https://api.github.com/users/nviraj/following{/other_user}", "gists_url": "https://api.github.com/users/nviraj/gists{/gist_id}", "starred_url": "https://api.github.com/users/nviraj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nviraj/subscriptions", "organizations_url": "https://api.github.com/users/nviraj/orgs", "repos_url": "https://api.github.com/users/nviraj/repos", "events_url": "https://api.github.com/users/nviraj/events{/privacy}", "received_events_url": "https://api.github.com/users/nviraj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
1
2024-08-01T12:25:31
2024-08-02T12:19:55
2024-08-02T12:19:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Apologies, if I am missing something but can you upload quants for [base 2b](https://huggingface.co/google/gemma-2-2b) as well? All I see are instruct ones [here](https://ollama.com/library/gemma2/tags) Thanks!
{ "login": "nviraj", "id": 8409854, "node_id": "MDQ6VXNlcjg0MDk4NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/8409854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nviraj", "html_url": "https://github.com/nviraj", "followers_url": "https://api.github.com/users/nviraj/followers", "following_url": "https://api.github.com/users/nviraj/following{/other_user}", "gists_url": "https://api.github.com/users/nviraj/gists{/gist_id}", "starred_url": "https://api.github.com/users/nviraj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nviraj/subscriptions", "organizations_url": "https://api.github.com/users/nviraj/orgs", "repos_url": "https://api.github.com/users/nviraj/repos", "events_url": "https://api.github.com/users/nviraj/events{/privacy}", "received_events_url": "https://api.github.com/users/nviraj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6117/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6117/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4652
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4652/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4652/comments
https://api.github.com/repos/ollama/ollama/issues/4652/events
https://github.com/ollama/ollama/pull/4652
2,317,888,554
PR_kwDOJ0Z1Ps5wmNS4
4,652
Ensure `nvidia` and `nvidia_uvm` kernel modules are loaded in `install.sh` script and at startup
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-26T18:53:36
2024-05-26T21:57:18
2024-05-26T21:57:17
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4652", "html_url": "https://github.com/ollama/ollama/pull/4652", "diff_url": "https://github.com/ollama/ollama/pull/4652.diff", "patch_url": "https://github.com/ollama/ollama/pull/4652.patch", "merged_at": "2024-05-26T21:57:17" }
Make sure the `nvidia` and `nvidia_uvm` kernel modules are loaded on install. Nvidia has a daemon that takes care of that `nvidia-persistenced` for restarts, so add both to its config file as well Lastly, make sure the kernel modules are loaded when re-running the install script, even if drivers are already installed. Fixes https://github.com/ollama/ollama/issues/4563
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4652/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4652/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1988
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1988/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1988/comments
https://api.github.com/repos/ollama/ollama/issues/1988/events
https://github.com/ollama/ollama/pull/1988
2,080,801,827
PR_kwDOJ0Z1Ps5kBthD
1,988
Fix typo in arm mac arch script
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-14T16:34:14
2024-01-14T16:45:25
2024-01-14T16:45:18
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1988", "html_url": "https://github.com/ollama/ollama/pull/1988", "diff_url": "https://github.com/ollama/ollama/pull/1988.diff", "patch_url": "https://github.com/ollama/ollama/pull/1988.patch", "merged_at": "2024-01-14T16:45:18" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1988/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1988/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3401
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3401/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3401/comments
https://api.github.com/repos/ollama/ollama/issues/3401/events
https://github.com/ollama/ollama/issues/3401
2,214,382,654
I_kwDOJ0Z1Ps6D_Mw-
3,401
Document `OLLAMA_DEBUG` in `ollama serve` `-h` docs
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5667396210, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2acg", "url": "https://api.github.com/repos/ollama/ollama/labels/good%20first%20issue", "name": "good first issue", "color": "7057ff", "default": true, "description": "Good for newcomers" } ]
closed
false
null
[]
null
2
2024-03-29T00:27:47
2024-05-13T20:20:58
2024-05-13T20:20:58
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? _No response_ ### How should we solve this? _No response_ ### What is the impact of not solving this? _No response_ ### Anything else? _No response_
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3401/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3401/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1648
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1648/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1648/comments
https://api.github.com/repos/ollama/ollama/issues/1648/events
https://github.com/ollama/ollama/issues/1648
2,051,491,566
I_kwDOJ0Z1Ps56R0bu
1,648
gpu issues
{ "login": "RootnuII", "id": 66104474, "node_id": "MDQ6VXNlcjY2MTA0NDc0", "avatar_url": "https://avatars.githubusercontent.com/u/66104474?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RootnuII", "html_url": "https://github.com/RootnuII", "followers_url": "https://api.github.com/users/RootnuII/followers", "following_url": "https://api.github.com/users/RootnuII/following{/other_user}", "gists_url": "https://api.github.com/users/RootnuII/gists{/gist_id}", "starred_url": "https://api.github.com/users/RootnuII/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RootnuII/subscriptions", "organizations_url": "https://api.github.com/users/RootnuII/orgs", "repos_url": "https://api.github.com/users/RootnuII/repos", "events_url": "https://api.github.com/users/RootnuII/events{/privacy}", "received_events_url": "https://api.github.com/users/RootnuII/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2023-12-21T00:38:19
2023-12-23T15:16:26
2023-12-23T09:59:42
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
warning: gpu support may not be enabled, check that you have installed GPU drivers: nvidia-smi command failed i have installed all drivers for my gpu:1070ti still nothing maybe i installed the wrong drivers can someone point me in the right direction
{ "login": "RootnuII", "id": 66104474, "node_id": "MDQ6VXNlcjY2MTA0NDc0", "avatar_url": "https://avatars.githubusercontent.com/u/66104474?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RootnuII", "html_url": "https://github.com/RootnuII", "followers_url": "https://api.github.com/users/RootnuII/followers", "following_url": "https://api.github.com/users/RootnuII/following{/other_user}", "gists_url": "https://api.github.com/users/RootnuII/gists{/gist_id}", "starred_url": "https://api.github.com/users/RootnuII/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RootnuII/subscriptions", "organizations_url": "https://api.github.com/users/RootnuII/orgs", "repos_url": "https://api.github.com/users/RootnuII/repos", "events_url": "https://api.github.com/users/RootnuII/events{/privacy}", "received_events_url": "https://api.github.com/users/RootnuII/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1648/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1648/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2828
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2828/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2828/comments
https://api.github.com/repos/ollama/ollama/issues/2828/events
https://github.com/ollama/ollama/pull/2828
2,160,642,796
PR_kwDOJ0Z1Ps5oQuDE
2,828
update llama.cpp submodule commit to `87c91c07`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-29T07:59:29
2024-02-29T17:42:09
2024-02-29T17:42:08
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2828", "html_url": "https://github.com/ollama/ollama/pull/2828", "diff_url": "https://github.com/ollama/ollama/pull/2828.diff", "patch_url": "https://github.com/ollama/ollama/pull/2828.patch", "merged_at": "2024-02-29T17:42:08" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2828/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2828/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1501
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1501/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1501/comments
https://api.github.com/repos/ollama/ollama/issues/1501/events
https://github.com/ollama/ollama/issues/1501
2,039,679,106
I_kwDOJ0Z1Ps55kwiC
1,501
macOS environment variable not working
{ "login": "brandoncarl", "id": 5427971, "node_id": "MDQ6VXNlcjU0Mjc5NzE=", "avatar_url": "https://avatars.githubusercontent.com/u/5427971?v=4", "gravatar_id": "", "url": "https://api.github.com/users/brandoncarl", "html_url": "https://github.com/brandoncarl", "followers_url": "https://api.github.com/users/brandoncarl/followers", "following_url": "https://api.github.com/users/brandoncarl/following{/other_user}", "gists_url": "https://api.github.com/users/brandoncarl/gists{/gist_id}", "starred_url": "https://api.github.com/users/brandoncarl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/brandoncarl/subscriptions", "organizations_url": "https://api.github.com/users/brandoncarl/orgs", "repos_url": "https://api.github.com/users/brandoncarl/repos", "events_url": "https://api.github.com/users/brandoncarl/events{/privacy}", "received_events_url": "https://api.github.com/users/brandoncarl/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
9
2023-12-13T13:08:29
2025-01-11T11:41:21
2024-01-22T23:41:41
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello - thanks for the great repository. I wanted to alert you to the fact that the OLLAMA_MODELS path appears to be having no impact. This is true for `pull`, `rull` and serving. ``` $ echo $OLLAMA_MODELS (prints appropriate directory) $ ollama run <model> (downloads to ~/.ollama/..) $ OLLAMA_MODELS=<directory> run <model> (downloads to ~/.ollama/...) ```
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1501/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1501/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4947
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4947/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4947/comments
https://api.github.com/repos/ollama/ollama/issues/4947/events
https://github.com/ollama/ollama/issues/4947
2,342,185,507
I_kwDOJ0Z1Ps6Lmuoj
4,947
Cannot update ollama on windows through "restart to update"
{ "login": "TonyBlur", "id": 101348912, "node_id": "U_kgDOBgp2MA", "avatar_url": "https://avatars.githubusercontent.com/u/101348912?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TonyBlur", "html_url": "https://github.com/TonyBlur", "followers_url": "https://api.github.com/users/TonyBlur/followers", "following_url": "https://api.github.com/users/TonyBlur/following{/other_user}", "gists_url": "https://api.github.com/users/TonyBlur/gists{/gist_id}", "starred_url": "https://api.github.com/users/TonyBlur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TonyBlur/subscriptions", "organizations_url": "https://api.github.com/users/TonyBlur/orgs", "repos_url": "https://api.github.com/users/TonyBlur/repos", "events_url": "https://api.github.com/users/TonyBlur/events{/privacy}", "received_events_url": "https://api.github.com/users/TonyBlur/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
11
2024-06-09T11:52:04
2024-06-10T11:42:01
2024-06-10T11:42:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? cannot update ollama on windows through clicking "restart to update" at taskbar, it has no response ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version 0.1.41
{ "login": "TonyBlur", "id": 101348912, "node_id": "U_kgDOBgp2MA", "avatar_url": "https://avatars.githubusercontent.com/u/101348912?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TonyBlur", "html_url": "https://github.com/TonyBlur", "followers_url": "https://api.github.com/users/TonyBlur/followers", "following_url": "https://api.github.com/users/TonyBlur/following{/other_user}", "gists_url": "https://api.github.com/users/TonyBlur/gists{/gist_id}", "starred_url": "https://api.github.com/users/TonyBlur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TonyBlur/subscriptions", "organizations_url": "https://api.github.com/users/TonyBlur/orgs", "repos_url": "https://api.github.com/users/TonyBlur/repos", "events_url": "https://api.github.com/users/TonyBlur/events{/privacy}", "received_events_url": "https://api.github.com/users/TonyBlur/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4947/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4947/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5918
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5918/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5918/comments
https://api.github.com/repos/ollama/ollama/issues/5918/events
https://github.com/ollama/ollama/issues/5918
2,427,947,507
I_kwDOJ0Z1Ps6Qt4nz
5,918
Llama3.1 70b-instruct-q4_1 buggy
{ "login": "velaia", "id": 1515904, "node_id": "MDQ6VXNlcjE1MTU5MDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1515904?v=4", "gravatar_id": "", "url": "https://api.github.com/users/velaia", "html_url": "https://github.com/velaia", "followers_url": "https://api.github.com/users/velaia/followers", "following_url": "https://api.github.com/users/velaia/following{/other_user}", "gists_url": "https://api.github.com/users/velaia/gists{/gist_id}", "starred_url": "https://api.github.com/users/velaia/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/velaia/subscriptions", "organizations_url": "https://api.github.com/users/velaia/orgs", "repos_url": "https://api.github.com/users/velaia/repos", "events_url": "https://api.github.com/users/velaia/events{/privacy}", "received_events_url": "https://api.github.com/users/velaia/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
14
2024-07-24T15:54:38
2024-09-15T17:35:22
2024-09-15T17:18:31
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When I run the **70b-instruct-q4_1** version of Llama3.1 ollama gives a buggy reply: My sample request: > ➜ ollama-tests curl http://localhost:11434/api/chat -d '{ "model": "llama3.1:70b-instruct-q4_1", "messages": [ { "role": "system", "content": "You are a helpful assistant." }, { "role": "user", "content": "Tell me about the top 3 commercial industrial computer vision products on the market." } ], "stream": false }' The model's response > { "model": "llama3.1:70b-instruct-q4_1", "created_at": "2024-07-24T15:45:23.026538Z", "message": { "role": "assistant", "content": "assistant\nassistantassistantassistant" }, "done_reason": "stop", "done": true, "total_duration": 56814340833, "load_duration": 53209095791, "prompt_eval_count": 37, "prompt_eval_duration": 2107466000, "eval_count": 6, "eval_duration": 1493717000 }` The same request generates a proper response using **llama3.1:8b-instruct-q8_0**. ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.2.8
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5918/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5918/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/2496
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2496/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2496/comments
https://api.github.com/repos/ollama/ollama/issues/2496/events
https://github.com/ollama/ollama/issues/2496
2,134,803,651
I_kwDOJ0Z1Ps5_PoTD
2,496
default num_thread incorrect on some large core count system (non-hyperthreading)
{ "login": "mokkin", "id": 2938748, "node_id": "MDQ6VXNlcjI5Mzg3NDg=", "avatar_url": "https://avatars.githubusercontent.com/u/2938748?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mokkin", "html_url": "https://github.com/mokkin", "followers_url": "https://api.github.com/users/mokkin/followers", "following_url": "https://api.github.com/users/mokkin/following{/other_user}", "gists_url": "https://api.github.com/users/mokkin/gists{/gist_id}", "starred_url": "https://api.github.com/users/mokkin/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mokkin/subscriptions", "organizations_url": "https://api.github.com/users/mokkin/orgs", "repos_url": "https://api.github.com/users/mokkin/repos", "events_url": "https://api.github.com/users/mokkin/events{/privacy}", "received_events_url": "https://api.github.com/users/mokkin/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
35
2024-02-14T17:10:39
2024-08-05T22:20:08
2024-08-05T22:20:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have tested Ollama on different machines yet, but no matter how many cores or RAM I have, it's only using 50% of the cores and just a very few GB of RAM. For example now I'm running `ollama rum llama2:70b` on 16 core server with 32 GB of RAM, but while prompting only eight cores are used and just around 1 GB of RAM. Is there something wrong? In the models descriptions are aleways warning you neet at least 8,16,32,... GB of RAM. ![Bildschirmfoto vom 2024-02-14 18-08-47](https://github.com/ollama/ollama/assets/2938748/8a47ec55-475d-4311-8110-3ca1e0a34cb8)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2496/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2496/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3754
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3754/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3754/comments
https://api.github.com/repos/ollama/ollama/issues/3754/events
https://github.com/ollama/ollama/issues/3754
2,253,202,923
I_kwDOJ0Z1Ps6GTSXr
3,754
ollama rm Support deletion through ID
{ "login": "eryajf", "id": 33259379, "node_id": "MDQ6VXNlcjMzMjU5Mzc5", "avatar_url": "https://avatars.githubusercontent.com/u/33259379?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eryajf", "html_url": "https://github.com/eryajf", "followers_url": "https://api.github.com/users/eryajf/followers", "following_url": "https://api.github.com/users/eryajf/following{/other_user}", "gists_url": "https://api.github.com/users/eryajf/gists{/gist_id}", "starred_url": "https://api.github.com/users/eryajf/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eryajf/subscriptions", "organizations_url": "https://api.github.com/users/eryajf/orgs", "repos_url": "https://api.github.com/users/eryajf/repos", "events_url": "https://api.github.com/users/eryajf/events{/privacy}", "received_events_url": "https://api.github.com/users/eryajf/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-04-19T14:48:40
2024-04-19T14:48:40
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Just like Docker RM supports deletion through ID, have you considered supporting model deletion through ID? If so, I am willing to research this area and submit a PR
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3754/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3754/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/225
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/225/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/225/comments
https://api.github.com/repos/ollama/ollama/issues/225/events
https://github.com/ollama/ollama/pull/225
1,824,949,943
PR_kwDOJ0Z1Ps5Wk5I4
225
add stop conditions
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-07-27T18:28:50
2023-07-28T00:20:57
2023-07-28T00:20:56
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/225", "html_url": "https://github.com/ollama/ollama/pull/225", "diff_url": "https://github.com/ollama/ollama/pull/225.diff", "patch_url": "https://github.com/ollama/ollama/pull/225.patch", "merged_at": "2023-07-28T00:20:56" }
resolves #140 resolves #217
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/225/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/225/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1653
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1653/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1653/comments
https://api.github.com/repos/ollama/ollama/issues/1653/events
https://github.com/ollama/ollama/issues/1653
2,052,245,322
I_kwDOJ0Z1Ps56UsdK
1,653
shell autocompletion
{ "login": "teto", "id": 886074, "node_id": "MDQ6VXNlcjg4NjA3NA==", "avatar_url": "https://avatars.githubusercontent.com/u/886074?v=4", "gravatar_id": "", "url": "https://api.github.com/users/teto", "html_url": "https://github.com/teto", "followers_url": "https://api.github.com/users/teto/followers", "following_url": "https://api.github.com/users/teto/following{/other_user}", "gists_url": "https://api.github.com/users/teto/gists{/gist_id}", "starred_url": "https://api.github.com/users/teto/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/teto/subscriptions", "organizations_url": "https://api.github.com/users/teto/orgs", "repos_url": "https://api.github.com/users/teto/repos", "events_url": "https://api.github.com/users/teto/events{/privacy}", "received_events_url": "https://api.github.com/users/teto/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
6
2023-12-21T11:28:48
2024-12-04T05:19:01
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
hi, I would like to add shell autocompletion for ollama to my linux distribution. I dont know go but the cli parser seems to be using https://github.com/spf13/cobra, the same as github CLI `gh`. The nix package generates the autocompletion like this: https://github.com/teto/nixpkgs/blob/dbaa028d61848fda534a04ce21459b722f0bca81/pkgs/applications/version-management/gh/default.nix#L30 so I tried `ollama completion -s zsh` hoping it was not shown by `ollama --help` because hidden but that failed. Is there a way to generate those files currently and if yes, how, if not, could you add such a subcommand please ? makes writing commands on the prompt easier.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1653/reactions", "total_count": 23, "+1": 23, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1653/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3365
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3365/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3365/comments
https://api.github.com/repos/ollama/ollama/issues/3365/events
https://github.com/ollama/ollama/issues/3365
2,209,739,153
I_kwDOJ0Z1Ps6DtfGR
3,365
Unable to make cors work in docker container
{ "login": "gbarton", "id": 1878953, "node_id": "MDQ6VXNlcjE4Nzg5NTM=", "avatar_url": "https://avatars.githubusercontent.com/u/1878953?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gbarton", "html_url": "https://github.com/gbarton", "followers_url": "https://api.github.com/users/gbarton/followers", "following_url": "https://api.github.com/users/gbarton/following{/other_user}", "gists_url": "https://api.github.com/users/gbarton/gists{/gist_id}", "starred_url": "https://api.github.com/users/gbarton/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gbarton/subscriptions", "organizations_url": "https://api.github.com/users/gbarton/orgs", "repos_url": "https://api.github.com/users/gbarton/repos", "events_url": "https://api.github.com/users/gbarton/events{/privacy}", "received_events_url": "https://api.github.com/users/gbarton/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
3
2024-03-27T03:24:08
2024-05-02T02:15:59
2024-03-27T22:44:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I have been trying to get cors to work inside a docker container with no luck so far. Attached is my docker-compose.yaml for starting ollama. I have tried many variations on the OLLAMA_HOST and OLLAMA_ORIGINS env vars to no luck. Does anyone know the magic combo to make it work? ### What did you expect to see? hitting it from a browser directly works fine, from code works fine, using Ollama chats https://github.com/drazdra/ollama-chats hosted on another node runs into cors fun: ![image](https://github.com/ollama/ollama/assets/1878953/fea70cec-b52a-441a-b163-d9bf23b6af99) curling on the local box shows anything set as origin is blocked: ```bash gman@ai:~/wrk/ai-runtime$ curl -i -H 'Origin:http://10.22.22.1' localhost:11434/ HTTP/1.1 403 Forbidden Date: Wed, 27 Mar 2024 03:20:24 GMT Content-Length: 0 gman@ai:~/wrk/ai-runtime$ curl localhost:11434 Ollama is running ``` ### Steps to reproduce docker-compose.yaml ``` version: '3.6' services: ollama: container_name: ollama deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] image: ollama/ollama #tty: true restart: unless-stopped environment: - OLLAMA_ORIGINS="*" - OLLAMA_HOST="0.0.0.0:11434" ports: - 11434:11434 volumes: - ./ollama:/root/.ollama:cached ``` ### Are there any recent changes that introduced the issue? _No response_ ### OS Linux ### Architecture amd64 ### Platform Docker ### Ollama version 0.1.29 ### GPU Nvidia ### GPU info _No response_ ### CPU AMD ### Other software _No response_
{ "login": "gbarton", "id": 1878953, "node_id": "MDQ6VXNlcjE4Nzg5NTM=", "avatar_url": "https://avatars.githubusercontent.com/u/1878953?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gbarton", "html_url": "https://github.com/gbarton", "followers_url": "https://api.github.com/users/gbarton/followers", "following_url": "https://api.github.com/users/gbarton/following{/other_user}", "gists_url": "https://api.github.com/users/gbarton/gists{/gist_id}", "starred_url": "https://api.github.com/users/gbarton/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gbarton/subscriptions", "organizations_url": "https://api.github.com/users/gbarton/orgs", "repos_url": "https://api.github.com/users/gbarton/repos", "events_url": "https://api.github.com/users/gbarton/events{/privacy}", "received_events_url": "https://api.github.com/users/gbarton/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3365/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3365/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2686
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2686/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2686/comments
https://api.github.com/repos/ollama/ollama/issues/2686/events
https://github.com/ollama/ollama/pull/2686
2,149,472,407
PR_kwDOJ0Z1Ps5nqz9L
2,686
update README to add Gemma 2B, 7B model in Model Library Table
{ "login": "adminazhar", "id": 20738252, "node_id": "MDQ6VXNlcjIwNzM4MjUy", "avatar_url": "https://avatars.githubusercontent.com/u/20738252?v=4", "gravatar_id": "", "url": "https://api.github.com/users/adminazhar", "html_url": "https://github.com/adminazhar", "followers_url": "https://api.github.com/users/adminazhar/followers", "following_url": "https://api.github.com/users/adminazhar/following{/other_user}", "gists_url": "https://api.github.com/users/adminazhar/gists{/gist_id}", "starred_url": "https://api.github.com/users/adminazhar/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/adminazhar/subscriptions", "organizations_url": "https://api.github.com/users/adminazhar/orgs", "repos_url": "https://api.github.com/users/adminazhar/repos", "events_url": "https://api.github.com/users/adminazhar/events{/privacy}", "received_events_url": "https://api.github.com/users/adminazhar/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-22T16:36:57
2024-02-22T18:07:47
2024-02-22T18:07:47
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2686", "html_url": "https://github.com/ollama/ollama/pull/2686", "diff_url": "https://github.com/ollama/ollama/pull/2686.diff", "patch_url": "https://github.com/ollama/ollama/pull/2686.patch", "merged_at": "2024-02-22T18:07:47" }
@pfrankov @jmorganca Kindly Review
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2686/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2686/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2911
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2911/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2911/comments
https://api.github.com/repos/ollama/ollama/issues/2911/events
https://github.com/ollama/ollama/issues/2911
2,166,737,391
I_kwDOJ0Z1Ps6BJcnv
2,911
System Monitor detected that internet is used by Ollama.
{ "login": "kukidevs", "id": 113847173, "node_id": "U_kgDOBskrhQ", "avatar_url": "https://avatars.githubusercontent.com/u/113847173?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kukidevs", "html_url": "https://github.com/kukidevs", "followers_url": "https://api.github.com/users/kukidevs/followers", "following_url": "https://api.github.com/users/kukidevs/following{/other_user}", "gists_url": "https://api.github.com/users/kukidevs/gists{/gist_id}", "starred_url": "https://api.github.com/users/kukidevs/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kukidevs/subscriptions", "organizations_url": "https://api.github.com/users/kukidevs/orgs", "repos_url": "https://api.github.com/users/kukidevs/repos", "events_url": "https://api.github.com/users/kukidevs/events{/privacy}", "received_events_url": "https://api.github.com/users/kukidevs/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-03-04T12:29:53
2025-01-30T01:41:17
2024-03-04T17:50:06
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
![image](https://github.com/ollama/ollama/assets/113847173/b90b5c11-2ce5-4f1f-a7ee-59abfff867d8) What does it send/receive? I was using it through MacOS terminal, installed the app from the website.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2911/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2911/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2200
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2200/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2200/comments
https://api.github.com/repos/ollama/ollama/issues/2200/events
https://github.com/ollama/ollama/issues/2200
2,101,705,603
I_kwDOJ0Z1Ps59RXuD
2,200
ollama is stuck when translate language
{ "login": "MunhakKim", "id": 13396311, "node_id": "MDQ6VXNlcjEzMzk2MzEx", "avatar_url": "https://avatars.githubusercontent.com/u/13396311?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MunhakKim", "html_url": "https://github.com/MunhakKim", "followers_url": "https://api.github.com/users/MunhakKim/followers", "following_url": "https://api.github.com/users/MunhakKim/following{/other_user}", "gists_url": "https://api.github.com/users/MunhakKim/gists{/gist_id}", "starred_url": "https://api.github.com/users/MunhakKim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MunhakKim/subscriptions", "organizations_url": "https://api.github.com/users/MunhakKim/orgs", "repos_url": "https://api.github.com/users/MunhakKim/repos", "events_url": "https://api.github.com/users/MunhakKim/events{/privacy}", "received_events_url": "https://api.github.com/users/MunhakKim/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
2
2024-01-26T07:10:48
2024-03-11T23:14:33
2024-03-11T23:14:33
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
ollama is stuck when i ask to translate language. In my case, i put it in like '한글로' for prompt which means 'in Korean' after English output. Sometimes ollama could translate perfectly and stable, but mostly ollama is stuck. I think if there is a problem in text generation, It should be pause not is broken. In this case, **i have to restart ollama**. I guess it is about string format from streaming. i am using docker image from docker hub. using this command 'docker exec -it ollama ollama run mistral' or api. version: v0.1.21 (i have been tested from v0.1.18~) model: mistral, mixtral latest, llama2 etc. sorry for no logs. But it is easy to cause problems when you put '한글로' in prompt.
{ "login": "MunhakKim", "id": 13396311, "node_id": "MDQ6VXNlcjEzMzk2MzEx", "avatar_url": "https://avatars.githubusercontent.com/u/13396311?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MunhakKim", "html_url": "https://github.com/MunhakKim", "followers_url": "https://api.github.com/users/MunhakKim/followers", "following_url": "https://api.github.com/users/MunhakKim/following{/other_user}", "gists_url": "https://api.github.com/users/MunhakKim/gists{/gist_id}", "starred_url": "https://api.github.com/users/MunhakKim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MunhakKim/subscriptions", "organizations_url": "https://api.github.com/users/MunhakKim/orgs", "repos_url": "https://api.github.com/users/MunhakKim/repos", "events_url": "https://api.github.com/users/MunhakKim/events{/privacy}", "received_events_url": "https://api.github.com/users/MunhakKim/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2200/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2200/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/870
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/870/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/870/comments
https://api.github.com/repos/ollama/ollama/issues/870/events
https://github.com/ollama/ollama/pull/870
1,955,433,858
PR_kwDOJ0Z1Ps5dcU35
870
Added a minimalist React UI for Ollama models to ReadME.md
{ "login": "richawo", "id": 35015261, "node_id": "MDQ6VXNlcjM1MDE1MjYx", "avatar_url": "https://avatars.githubusercontent.com/u/35015261?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richawo", "html_url": "https://github.com/richawo", "followers_url": "https://api.github.com/users/richawo/followers", "following_url": "https://api.github.com/users/richawo/following{/other_user}", "gists_url": "https://api.github.com/users/richawo/gists{/gist_id}", "starred_url": "https://api.github.com/users/richawo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/richawo/subscriptions", "organizations_url": "https://api.github.com/users/richawo/orgs", "repos_url": "https://api.github.com/users/richawo/repos", "events_url": "https://api.github.com/users/richawo/events{/privacy}", "received_events_url": "https://api.github.com/users/richawo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2023-10-21T10:50:25
2023-10-23T14:44:39
2023-10-23T14:44:39
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/870", "html_url": "https://github.com/ollama/ollama/pull/870", "diff_url": "https://github.com/ollama/ollama/pull/870.diff", "patch_url": "https://github.com/ollama/ollama/pull/870.patch", "merged_at": "2023-10-23T14:44:39" }
I discussed it a few times in the discord, and a few people seem to be using it, so it would be good to add. Demo video: https://github.com/jmorganca/ollama/assets/35015261/d50f7036-cdf2-44ed-9bb0-fdbed6a4ec66 I'll be maintaining/improving it a lot over the coming weeks, and some contributors reached out to get involved. It can now handle markdown, etc. And continues to improve beyond the initial demo.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/870/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/870/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2587
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2587/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2587/comments
https://api.github.com/repos/ollama/ollama/issues/2587/events
https://github.com/ollama/ollama/issues/2587
2,141,539,255
I_kwDOJ0Z1Ps5_pUu3
2,587
Running on GPU
{ "login": "shersoni610", "id": 57876250, "node_id": "MDQ6VXNlcjU3ODc2MjUw", "avatar_url": "https://avatars.githubusercontent.com/u/57876250?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shersoni610", "html_url": "https://github.com/shersoni610", "followers_url": "https://api.github.com/users/shersoni610/followers", "following_url": "https://api.github.com/users/shersoni610/following{/other_user}", "gists_url": "https://api.github.com/users/shersoni610/gists{/gist_id}", "starred_url": "https://api.github.com/users/shersoni610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shersoni610/subscriptions", "organizations_url": "https://api.github.com/users/shersoni610/orgs", "repos_url": "https://api.github.com/users/shersoni610/repos", "events_url": "https://api.github.com/users/shersoni610/events{/privacy}", "received_events_url": "https://api.github.com/users/shersoni610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
29
2024-02-19T05:27:39
2024-02-29T23:47:57
2024-02-29T23:47:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello, It seems, the response time of llama2:7b is slow on my linux machine. I am not sure if the code is running on Nvidia card. In a python code, how to ensure that Ollama models run on GPU?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2587/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2587/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5742
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5742/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5742/comments
https://api.github.com/repos/ollama/ollama/issues/5742/events
https://github.com/ollama/ollama/pull/5742
2,413,170,315
PR_kwDOJ0Z1Ps51nl1k
5,742
Added code to use swap memory in linux
{ "login": "avinash-ghadshi", "id": 65444549, "node_id": "MDQ6VXNlcjY1NDQ0NTQ5", "avatar_url": "https://avatars.githubusercontent.com/u/65444549?v=4", "gravatar_id": "", "url": "https://api.github.com/users/avinash-ghadshi", "html_url": "https://github.com/avinash-ghadshi", "followers_url": "https://api.github.com/users/avinash-ghadshi/followers", "following_url": "https://api.github.com/users/avinash-ghadshi/following{/other_user}", "gists_url": "https://api.github.com/users/avinash-ghadshi/gists{/gist_id}", "starred_url": "https://api.github.com/users/avinash-ghadshi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/avinash-ghadshi/subscriptions", "organizations_url": "https://api.github.com/users/avinash-ghadshi/orgs", "repos_url": "https://api.github.com/users/avinash-ghadshi/repos", "events_url": "https://api.github.com/users/avinash-ghadshi/events{/privacy}", "received_events_url": "https://api.github.com/users/avinash-ghadshi/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2024-07-17T09:54:10
2024-09-03T16:54:39
2024-09-03T16:54:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5742", "html_url": "https://github.com/ollama/ollama/pull/5742", "diff_url": "https://github.com/ollama/ollama/pull/5742.diff", "patch_url": "https://github.com/ollama/ollama/pull/5742.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5742/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5742/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6420
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6420/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6420/comments
https://api.github.com/repos/ollama/ollama/issues/6420/events
https://github.com/ollama/ollama/issues/6420
2,473,039,397
I_kwDOJ0Z1Ps6TZ5Yl
6,420
Is the speed of the Olama running model related to the CUDA version?
{ "login": "TianWuYuJiangHenShou", "id": 20592000, "node_id": "MDQ6VXNlcjIwNTkyMDAw", "avatar_url": "https://avatars.githubusercontent.com/u/20592000?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TianWuYuJiangHenShou", "html_url": "https://github.com/TianWuYuJiangHenShou", "followers_url": "https://api.github.com/users/TianWuYuJiangHenShou/followers", "following_url": "https://api.github.com/users/TianWuYuJiangHenShou/following{/other_user}", "gists_url": "https://api.github.com/users/TianWuYuJiangHenShou/gists{/gist_id}", "starred_url": "https://api.github.com/users/TianWuYuJiangHenShou/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TianWuYuJiangHenShou/subscriptions", "organizations_url": "https://api.github.com/users/TianWuYuJiangHenShou/orgs", "repos_url": "https://api.github.com/users/TianWuYuJiangHenShou/repos", "events_url": "https://api.github.com/users/TianWuYuJiangHenShou/events{/privacy}", "received_events_url": "https://api.github.com/users/TianWuYuJiangHenShou/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" }, { "id": 6677745918, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g", "url": "https://api.github.com/repos/ollama/ollama/labels/gpu", "name": "gpu", "color": "76C49E", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-08-19T10:40:53
2024-09-26T00:13:40
2024-09-26T00:13:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I deployed qwen2:72B with the latest version of Ollama, but I found that the loading speed of Ollama models varies greatly under different nvidia driver version. # driver:535.183.06 | cuda version:12.2 ollama version:0.3.4 Time of Loading Model:29s # driver:515.105.01 | cuda version:11.7 ollama version:0.3.6 Time of Loading Model:659s ## GPU :A800 ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version _No response_
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6420/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6420/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1418
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1418/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1418/comments
https://api.github.com/repos/ollama/ollama/issues/1418/events
https://github.com/ollama/ollama/issues/1418
2,031,151,545
I_kwDOJ0Z1Ps55EOm5
1,418
Make `notus` model available on `ollama`
{ "login": "davidberenstein1957", "id": 25269220, "node_id": "MDQ6VXNlcjI1MjY5MjIw", "avatar_url": "https://avatars.githubusercontent.com/u/25269220?v=4", "gravatar_id": "", "url": "https://api.github.com/users/davidberenstein1957", "html_url": "https://github.com/davidberenstein1957", "followers_url": "https://api.github.com/users/davidberenstein1957/followers", "following_url": "https://api.github.com/users/davidberenstein1957/following{/other_user}", "gists_url": "https://api.github.com/users/davidberenstein1957/gists{/gist_id}", "starred_url": "https://api.github.com/users/davidberenstein1957/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/davidberenstein1957/subscriptions", "organizations_url": "https://api.github.com/users/davidberenstein1957/orgs", "repos_url": "https://api.github.com/users/davidberenstein1957/repos", "events_url": "https://api.github.com/users/davidberenstein1957/events{/privacy}", "received_events_url": "https://api.github.com/users/davidberenstein1957/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
13
2023-12-07T16:53:09
2024-01-11T07:03:57
2024-01-11T07:03:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
We just hosted our [Argilla community meetup](https://www.youtube.com/playlist?list=PLBmuFBJ5cjcbsr49KFoC4DQoo3ZWT7q_d) and got the question if we would be able to make [Notus](https://argilla.io/blog/notus7b/) available on ollama. What would it take, and how can we help with this?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1418/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1418/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1693
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1693/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1693/comments
https://api.github.com/repos/ollama/ollama/issues/1693/events
https://github.com/ollama/ollama/issues/1693
2,054,975,190
I_kwDOJ0Z1Ps56fG7W
1,693
Possible to increase speed / efficiency of model?
{ "login": "theyluvEnething", "id": 115583828, "node_id": "U_kgDOBuOrVA", "avatar_url": "https://avatars.githubusercontent.com/u/115583828?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theyluvEnething", "html_url": "https://github.com/theyluvEnething", "followers_url": "https://api.github.com/users/theyluvEnething/followers", "following_url": "https://api.github.com/users/theyluvEnething/following{/other_user}", "gists_url": "https://api.github.com/users/theyluvEnething/gists{/gist_id}", "starred_url": "https://api.github.com/users/theyluvEnething/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/theyluvEnething/subscriptions", "organizations_url": "https://api.github.com/users/theyluvEnething/orgs", "repos_url": "https://api.github.com/users/theyluvEnething/repos", "events_url": "https://api.github.com/users/theyluvEnething/events{/privacy}", "received_events_url": "https://api.github.com/users/theyluvEnething/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2023-12-24T03:49:05
2024-06-22T13:47:35
2024-03-12T00:16:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm trying out the Dolphin-Mixture model and it's quite fun, but really slow. (my specs are 64gb 3200mj ram, i7 4.5gh cpu and a 1080 ti) but still it takes some time to start answering and when it starts it writes at maybe 2-3 words a second. Is it possible to make improvements to this?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1693/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1693/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4172
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4172/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4172/comments
https://api.github.com/repos/ollama/ollama/issues/4172/events
https://github.com/ollama/ollama/pull/4172
2,279,565,700
PR_kwDOJ0Z1Ps5ukaTU
4,172
Add support for the BSD platforms
{ "login": "prep", "id": 2017915, "node_id": "MDQ6VXNlcjIwMTc5MTU=", "avatar_url": "https://avatars.githubusercontent.com/u/2017915?v=4", "gravatar_id": "", "url": "https://api.github.com/users/prep", "html_url": "https://github.com/prep", "followers_url": "https://api.github.com/users/prep/followers", "following_url": "https://api.github.com/users/prep/following{/other_user}", "gists_url": "https://api.github.com/users/prep/gists{/gist_id}", "starred_url": "https://api.github.com/users/prep/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/prep/subscriptions", "organizations_url": "https://api.github.com/users/prep/orgs", "repos_url": "https://api.github.com/users/prep/repos", "events_url": "https://api.github.com/users/prep/events{/privacy}", "received_events_url": "https://api.github.com/users/prep/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
6
2024-05-05T14:42:06
2024-09-08T12:35:59
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4172", "html_url": "https://github.com/ollama/ollama/pull/4172", "diff_url": "https://github.com/ollama/ollama/pull/4172.diff", "patch_url": "https://github.com/ollama/ollama/pull/4172.patch", "merged_at": null }
This PR adds the ability to compile and run Ollama on various BSD platforms, specifically DragonFly, FreeBSD, NetBSD and OpenBSD. A couple notes: 1. It includes support for Vulkan to offload work to the GPU. 2. It's only been tested on OpenBSD, but hopefully interested parties in this community can test the other platforms. 3. `ollama run` is broken on OpenBSD. The [readline](readline) package uses `syscall` which OpenBSD [has removed from libc and the kernel](https://marc.info/?l=openbsd-tech&m=169841790407370&w=2). Instead, the readline package should be updated to use `golang.org/x/sys/unix`, similar to how [ergochat/readline](https://github.com/ergochat/readline) does it. I didn't want to include that work in this PR. 4. The tweak in `gen_common.sh` is to remove a bash'ism so that the file can be sourced from `ksh`. On OpenBSD I needed to add these packages: ```pkg_add cmake go vulkan-headers``` The other platforms will probably have similarly named packages to install.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4172/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4172/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8169
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8169/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8169/comments
https://api.github.com/repos/ollama/ollama/issues/8169/events
https://github.com/ollama/ollama/pull/8169
2,749,784,967
PR_kwDOJ0Z1Ps6FwPej
8,169
Update go.mod
{ "login": "Squishedmac", "id": 88924339, "node_id": "MDQ6VXNlcjg4OTI0MzM5", "avatar_url": "https://avatars.githubusercontent.com/u/88924339?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Squishedmac", "html_url": "https://github.com/Squishedmac", "followers_url": "https://api.github.com/users/Squishedmac/followers", "following_url": "https://api.github.com/users/Squishedmac/following{/other_user}", "gists_url": "https://api.github.com/users/Squishedmac/gists{/gist_id}", "starred_url": "https://api.github.com/users/Squishedmac/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Squishedmac/subscriptions", "organizations_url": "https://api.github.com/users/Squishedmac/orgs", "repos_url": "https://api.github.com/users/Squishedmac/repos", "events_url": "https://api.github.com/users/Squishedmac/events{/privacy}", "received_events_url": "https://api.github.com/users/Squishedmac/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-12-19T09:50:26
2024-12-19T10:51:33
2024-12-19T10:51:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8169", "html_url": "https://github.com/ollama/ollama/pull/8169", "diff_url": "https://github.com/ollama/ollama/pull/8169.diff", "patch_url": "https://github.com/ollama/ollama/pull/8169.patch", "merged_at": null }
Fixed the CVE-2024-45337 severity on wiz scan for the docker image, changing the version of golang.org/x/crypto to the latest version should fix the issue
{ "login": "Squishedmac", "id": 88924339, "node_id": "MDQ6VXNlcjg4OTI0MzM5", "avatar_url": "https://avatars.githubusercontent.com/u/88924339?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Squishedmac", "html_url": "https://github.com/Squishedmac", "followers_url": "https://api.github.com/users/Squishedmac/followers", "following_url": "https://api.github.com/users/Squishedmac/following{/other_user}", "gists_url": "https://api.github.com/users/Squishedmac/gists{/gist_id}", "starred_url": "https://api.github.com/users/Squishedmac/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Squishedmac/subscriptions", "organizations_url": "https://api.github.com/users/Squishedmac/orgs", "repos_url": "https://api.github.com/users/Squishedmac/repos", "events_url": "https://api.github.com/users/Squishedmac/events{/privacy}", "received_events_url": "https://api.github.com/users/Squishedmac/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8169/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8169/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6890
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6890/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6890/comments
https://api.github.com/repos/ollama/ollama/issues/6890/events
https://github.com/ollama/ollama/issues/6890
2,538,032,802
I_kwDOJ0Z1Ps6XR06i
6,890
"/show parameters" command causes crashes when running Qwen 2.5 models, on version 0.3.11
{ "login": "cmdntfnd", "id": 142593649, "node_id": "U_kgDOCH_OcQ", "avatar_url": "https://avatars.githubusercontent.com/u/142593649?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cmdntfnd", "html_url": "https://github.com/cmdntfnd", "followers_url": "https://api.github.com/users/cmdntfnd/followers", "following_url": "https://api.github.com/users/cmdntfnd/following{/other_user}", "gists_url": "https://api.github.com/users/cmdntfnd/gists{/gist_id}", "starred_url": "https://api.github.com/users/cmdntfnd/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cmdntfnd/subscriptions", "organizations_url": "https://api.github.com/users/cmdntfnd/orgs", "repos_url": "https://api.github.com/users/cmdntfnd/repos", "events_url": "https://api.github.com/users/cmdntfnd/events{/privacy}", "received_events_url": "https://api.github.com/users/cmdntfnd/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2024-09-20T06:54:55
2024-12-13T17:07:10
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? This only happens after changing the parameters through /set parameter command. Here's an example: PS H:\ztmp> ollama run qwen2.5 >>> /show parameters No parameters were specified for this model. >>> /set parameter top_k 1 Set parameter 'top_k' to '1' >>> /show parameters error: couldn't get model Error: something went wrong, please see the ollama server logs for details PS H:\ztmp> This is on windows 11. Here's the error message from the ollama serve terminal tab: 2024/09/20 08:47:43 [Recovery] 2024/09/20 - 08:47:43 panic recovered: assignment to entry in nil map runtime/map_faststr.go:205 (0x7a93ba) github.com/ollama/ollama/server/routes.go:807 (0x12cc57e) github.com/ollama/ollama/server/routes.go:732 (0x12cb497) github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1287cca) github.com/ollama/ollama/server/routes.go:1076 (0x12d0c14) github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1295d39) github.com/gin-gonic/gin@v1.10.0/recovery.go:102 (0x1295d27) github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1294e64) github.com/gin-gonic/gin@v1.10.0/logger.go:249 (0x1294e4b) github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1294291) github.com/gin-gonic/gin@v1.10.0/gin.go:633 (0x1293d00) github.com/gin-gonic/gin@v1.10.0/gin.go:589 (0x1293831) net/http/server.go:2688 (0xafaecc) net/http/server.go:3142 (0xafc6cd) net/http/server.go:2044 (0xaf79c7) runtime/asm_amd64.s:1695 (0x8026a0) ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.3.11
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6890/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6890/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1020
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1020/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1020/comments
https://api.github.com/repos/ollama/ollama/issues/1020/events
https://github.com/ollama/ollama/pull/1020
1,979,730,196
PR_kwDOJ0Z1Ps5euED0
1,020
Add ModelFusion community integration
{ "login": "lgrammel", "id": 205036, "node_id": "MDQ6VXNlcjIwNTAzNg==", "avatar_url": "https://avatars.githubusercontent.com/u/205036?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lgrammel", "html_url": "https://github.com/lgrammel", "followers_url": "https://api.github.com/users/lgrammel/followers", "following_url": "https://api.github.com/users/lgrammel/following{/other_user}", "gists_url": "https://api.github.com/users/lgrammel/gists{/gist_id}", "starred_url": "https://api.github.com/users/lgrammel/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lgrammel/subscriptions", "organizations_url": "https://api.github.com/users/lgrammel/orgs", "repos_url": "https://api.github.com/users/lgrammel/repos", "events_url": "https://api.github.com/users/lgrammel/events{/privacy}", "received_events_url": "https://api.github.com/users/lgrammel/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-11-06T17:54:56
2023-11-06T18:46:16
2023-11-06T18:46:16
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1020", "html_url": "https://github.com/ollama/ollama/pull/1020", "diff_url": "https://github.com/ollama/ollama/pull/1020.diff", "patch_url": "https://github.com/ollama/ollama/pull/1020.patch", "merged_at": "2023-11-06T18:46:16" }
null
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1020/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1020/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5656
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5656/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5656/comments
https://api.github.com/repos/ollama/ollama/issues/5656/events
https://github.com/ollama/ollama/pull/5656
2,406,435,963
PR_kwDOJ0Z1Ps51RF2g
5,656
llm: reorder gguf tensors
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-12T22:38:11
2024-08-14T16:49:46
2024-08-14T16:49:45
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5656", "html_url": "https://github.com/ollama/ollama/pull/5656", "diff_url": "https://github.com/ollama/ollama/pull/5656.diff", "patch_url": "https://github.com/ollama/ollama/pull/5656.patch", "merged_at": null }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5656/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5656/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5629
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5629/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5629/comments
https://api.github.com/repos/ollama/ollama/issues/5629/events
https://github.com/ollama/ollama/issues/5629
2,403,190,147
I_kwDOJ0Z1Ps6PPcWD
5,629
Crashing or gibberish output on 3x Radeon GPUs
{ "login": "darwinvelez58", "id": 118543481, "node_id": "U_kgDOBxDUeQ", "avatar_url": "https://avatars.githubusercontent.com/u/118543481?v=4", "gravatar_id": "", "url": "https://api.github.com/users/darwinvelez58", "html_url": "https://github.com/darwinvelez58", "followers_url": "https://api.github.com/users/darwinvelez58/followers", "following_url": "https://api.github.com/users/darwinvelez58/following{/other_user}", "gists_url": "https://api.github.com/users/darwinvelez58/gists{/gist_id}", "starred_url": "https://api.github.com/users/darwinvelez58/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/darwinvelez58/subscriptions", "organizations_url": "https://api.github.com/users/darwinvelez58/orgs", "repos_url": "https://api.github.com/users/darwinvelez58/repos", "events_url": "https://api.github.com/users/darwinvelez58/events{/privacy}", "received_events_url": "https://api.github.com/users/darwinvelez58/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6433346500, "node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA", "url": "https://api.github.com/repos/ollama/ollama/labels/amd", "name": "amd", "color": "000000", "default": false, "description": "Issues relating to AMD GPUs and ROCm" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
22
2024-07-11T13:26:43
2024-11-27T22:33:13
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? While running mixtral:8x7b-instruct-v0.1-q4_K_M on my physical machine with x3 7900 XTX I got this error: ``` [root@5dc6ecf27031 /]# ollama run mixtral:8x7b-instruct-v0.1-q4_K_M Error: llama runner process has terminated: signal: segmentation fault (core dumped) [root@5dc6ecf27031 /]# ``` Logs: ``` [GIN] 2024/07/11 - 13:22:44 | 200 | 16.23µs | 127.0.0.1 | HEAD "/" [GIN] 2024/07/11 - 13:22:44 | 200 | 7.724554ms | 127.0.0.1 | POST "/api/show" time=2024-07-11T13:22:44.297Z level=INFO source=sched.go:754 msg="new model will fit in available VRAM, loading" model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 library=rocm parallel=4 required="33.2 GiB" time=2024-07-11T13:22:44.298Z level=INFO source=memory.go:309 msg="offload to rocm" layers.requested=-1 layers.model=33 layers.offload=33 layers.split=11,11,11 memory.available="[24.0 GiB 24.0 GiB 24.0 GiB]" memory.required.full="33.2 GiB" memory.required.partial="33.2 GiB" memory.required.kv="1.0 GiB" memory.required.allocations="[11.3 GiB 11.3 GiB 10.6 GiB]" memory.weights.total="25.5 GiB" memory.weights.repeating="25.4 GiB" memory.weights.nonrepeating="102.6 MiB" memory.graph.full="1.3 GiB" memory.graph.partial="1.3 GiB" time=2024-07-11T13:22:44.299Z level=INFO source=server.go:375 msg="starting llama server" cmd="/tmp/ollama1419561683/runners/rocm_v60101/ollama_llama_server --model /root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 33 --parallel 4 --tensor-split 11,11,11 --tensor-split 11,11,11 --port 41695" time=2024-07-11T13:22:44.299Z level=INFO source=sched.go:474 msg="loaded runners" count=1 time=2024-07-11T13:22:44.299Z level=INFO source=server.go:563 msg="waiting for llama runner to start responding" time=2024-07-11T13:22:44.299Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server error" INFO [main] build info | build=1 commit="a8db2a9" tid="140134008951616" timestamp=1720704164 INFO [main] system info | n_threads=16 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 0 | " tid="140134008951616" timestamp=1720704164 total_threads=32 INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="31" port="41695" tid="140134008951616" timestamp=1720704164 llama_model_loader: loaded meta data with 26 key-value pairs and 995 tensors from /root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = llama llama_model_loader: - kv 1: general.name str = mistralai llama_model_loader: - kv 2: llama.context_length u32 = 32768 llama_model_loader: - kv 3: llama.embedding_length u32 = 4096 llama_model_loader: - kv 4: llama.block_count u32 = 32 llama_model_loader: - kv 5: llama.feed_forward_length u32 = 14336 llama_model_loader: - kv 6: llama.rope.dimension_count u32 = 128 llama_model_loader: - kv 7: llama.attention.head_count u32 = 32 llama_model_loader: - kv 8: llama.attention.head_count_kv u32 = 8 llama_model_loader: - kv 9: llama.expert_count u32 = 8 llama_model_loader: - kv 10: llama.expert_used_count u32 = 2 llama_model_loader: - kv 11: llama.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 12: llama.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 13: general.file_type u32 = 15 llama_model_loader: - kv 14: tokenizer.ggml.model str = llama llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,32000] = ["<unk>", "<s>", "</s>", "<0x00>", "<... llama_model_loader: - kv 16: tokenizer.ggml.scores arr[f32,32000] = [0.000000, 0.000000, 0.000000, 0.0000... llama_model_loader: - kv 17: tokenizer.ggml.token_type arr[i32,32000] = [2, 3, 3, 6, 6, 6, 6, 6, 6, 6, 6, 6, ... llama_model_loader: - kv 18: tokenizer.ggml.merges arr[str,58980] = ["▁ t", "i n", "e r", "▁ a", "h e... llama_model_loader: - kv 19: tokenizer.ggml.bos_token_id u32 = 1 llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 2 llama_model_loader: - kv 21: tokenizer.ggml.unknown_token_id u32 = 0 llama_model_loader: - kv 22: tokenizer.ggml.add_bos_token bool = true llama_model_loader: - kv 23: tokenizer.ggml.add_eos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {{ bos_token }}{% for message in mess... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - type f32: 65 tensors llama_model_loader: - type f16: 32 tensors llama_model_loader: - type q8_0: 64 tensors llama_model_loader: - type q4_K: 833 tensors llama_model_loader: - type q6_K: 1 tensors llm_load_vocab: special tokens cache size = 259 llm_load_vocab: token to piece cache size = 0.1637 MB llm_load_print_meta: format = GGUF V3 (latest) llm_load_print_meta: arch = llama llm_load_print_meta: vocab type = SPM llm_load_print_meta: n_vocab = 32000 llm_load_print_meta: n_merges = 0 llm_load_print_meta: vocab_only = 0 llm_load_print_meta: n_ctx_train = 32768 llm_load_print_meta: n_embd = 4096 llm_load_print_meta: n_layer = 32 llm_load_print_meta: n_head = 32 llm_load_print_meta: n_head_kv = 8 llm_load_print_meta: n_rot = 128 llm_load_print_meta: n_swa = 0 llm_load_print_meta: n_embd_head_k = 128 llm_load_print_meta: n_embd_head_v = 128 llm_load_print_meta: n_gqa = 4 llm_load_print_meta: n_embd_k_gqa = 1024 llm_load_print_meta: n_embd_v_gqa = 1024 llm_load_print_meta: f_norm_eps = 0.0e+00 llm_load_print_meta: f_norm_rms_eps = 1.0e-05 llm_load_print_meta: f_clamp_kqv = 0.0e+00 llm_load_print_meta: f_max_alibi_bias = 0.0e+00 llm_load_print_meta: f_logit_scale = 0.0e+00 llm_load_print_meta: n_ff = 14336 llm_load_print_meta: n_expert = 8 llm_load_print_meta: n_expert_used = 2 llm_load_print_meta: causal attn = 1 llm_load_print_meta: pooling type = 0 llm_load_print_meta: rope type = 0 llm_load_print_meta: rope scaling = linear llm_load_print_meta: freq_base_train = 1000000.0 llm_load_print_meta: freq_scale_train = 1 llm_load_print_meta: n_ctx_orig_yarn = 32768 llm_load_print_meta: rope_finetuned = unknown llm_load_print_meta: ssm_d_conv = 0 llm_load_print_meta: ssm_d_inner = 0 llm_load_print_meta: ssm_d_state = 0 llm_load_print_meta: ssm_dt_rank = 0 llm_load_print_meta: model type = 8x7B llm_load_print_meta: model ftype = Q4_K - Medium llm_load_print_meta: model params = 46.70 B llm_load_print_meta: model size = 24.62 GiB (4.53 BPW) llm_load_print_meta: general.name = mistralai llm_load_print_meta: BOS token = 1 '<s>' llm_load_print_meta: EOS token = 2 '</s>' llm_load_print_meta: UNK token = 0 '<unk>' llm_load_print_meta: LF token = 13 '<0x0A>' llm_load_print_meta: max token length = 48 time=2024-07-11T13:22:44.549Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server loading model" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 ROCm devices: Device 0: Radeon RX 7900 XTX, compute capability 11.0, VMM: no Device 1: Radeon RX 7900 XTX, compute capability 11.0, VMM: no Device 2: Radeon RX 7900 XTX, compute capability 11.0, VMM: no llm_load_tensors: ggml ctx size = 1.53 MiB llm_load_tensors: offloading 32 repeating layers to GPU llm_load_tensors: offloading non-repeating layers to GPU llm_load_tensors: offloaded 33/33 layers to GPU llm_load_tensors: ROCm0 buffer size = 8608.53 MiB llm_load_tensors: ROCm1 buffer size = 8608.53 MiB llm_load_tensors: ROCm2 buffer size = 7928.49 MiB llm_load_tensors: ROCm_Host buffer size = 70.31 MiB time=2024-07-11T13:23:03.566Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server not responding" llama_new_context_with_model: n_ctx = 8192 llama_new_context_with_model: n_batch = 512 llama_new_context_with_model: n_ubatch = 512 llama_new_context_with_model: flash_attn = 0 llama_new_context_with_model: freq_base = 1000000.0 llama_new_context_with_model: freq_scale = 1 time=2024-07-11T13:23:04.460Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server loading model" llama_kv_cache_init: ROCm0 KV buffer size = 352.00 MiB llama_kv_cache_init: ROCm1 KV buffer size = 352.00 MiB llama_kv_cache_init: ROCm2 KV buffer size = 320.00 MiB llama_new_context_with_model: KV self size = 1024.00 MiB, K (f16): 512.00 MiB, V (f16): 512.00 MiB llama_new_context_with_model: ROCm_Host output buffer size = 0.55 MiB llama_new_context_with_model: pipeline parallelism enabled (n_copies=4) llama_new_context_with_model: ROCm0 compute buffer size = 640.01 MiB llama_new_context_with_model: ROCm1 compute buffer size = 640.01 MiB llama_new_context_with_model: ROCm2 compute buffer size = 640.02 MiB llama_new_context_with_model: ROCm_Host compute buffer size = 72.02 MiB llama_new_context_with_model: graph nodes = 1510 llama_new_context_with_model: graph splits = 4 time=2024-07-11T13:23:06.864Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server error" [GIN] 2024/07/11 - 13:23:07 | 500 | 22.834580361s | 127.0.0.1 | POST "/api/chat" time=2024-07-11T13:23:07.115Z level=ERROR source=sched.go:480 msg="error loading llama server" error="llama runner process has terminated: signal: segmentation fault (core dumped) " time=2024-07-11T13:23:12.116Z level=WARN source=sched.go:671 msg="gpu VRAM usage didn't recover within timeout" seconds=5.001085328 model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 time=2024-07-11T13:23:12.366Z level=WARN source=sched.go:671 msg="gpu VRAM usage didn't recover within timeout" seconds=5.251122065 model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 time=2024-07-11T13:23:12.616Z level=WARN source=sched.go:671 msg="gpu VRAM usage didn't recover within timeout" seconds=5.500799906 model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 ``` I am running this docker version docker run -d --restart unless-stopped --device /dev/kfd --device /dev/dri -v ollama:/root/.ollama -p 11442:11434 --name dvz3 ollama/ollama:0.2.1-rocm ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.2.1-rocm
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5629/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5629/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1707
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1707/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1707/comments
https://api.github.com/repos/ollama/ollama/issues/1707/events
https://github.com/ollama/ollama/issues/1707
2,055,634,785
I_kwDOJ0Z1Ps56hn9h
1,707
[Bug] Calling embedding endpoint within short time will receive `None`
{ "login": "samx81", "id": 20315145, "node_id": "MDQ6VXNlcjIwMzE1MTQ1", "avatar_url": "https://avatars.githubusercontent.com/u/20315145?v=4", "gravatar_id": "", "url": "https://api.github.com/users/samx81", "html_url": "https://github.com/samx81", "followers_url": "https://api.github.com/users/samx81/followers", "following_url": "https://api.github.com/users/samx81/following{/other_user}", "gists_url": "https://api.github.com/users/samx81/gists{/gist_id}", "starred_url": "https://api.github.com/users/samx81/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/samx81/subscriptions", "organizations_url": "https://api.github.com/users/samx81/orgs", "repos_url": "https://api.github.com/users/samx81/repos", "events_url": "https://api.github.com/users/samx81/events{/privacy}", "received_events_url": "https://api.github.com/users/samx81/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2023-12-25T11:44:12
2024-05-06T23:44:40
2024-05-06T23:44:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm currently using `llama_index` to make LLM do a document QA, but I notice the embedding endpoint often output: ```json {"embedding":null} ``` when I do `condense_question` chat mode which calls LLM to rephrase question (and match context) before actually answer. Like: ```python llm = Ollama(model="dolphin2.2-mistral:7b-q4_K_M") ollama_embedding = OllamaEmbedding( model_name="dolphin2.2-mistral:7b-q4_K_M", ollama_additional_kwargs={"mirostat": 0}, ) service_context = ServiceContext.from_defaults(llm=llm, embed_model=ollama_embedding) chat_engine = index.as_chat_engine(service_context=service_context,chat_mode='condense_plus_context') resp= chat_engine.chat('abc') print(resp) # Often happens when LLM is called second time resp= chat_engine.chat('cde') print(resp) ``` I guess it is probably due to ollama or llama.cpp needs to handle different behavior in the same time? Because if I modify `llama_index` to check for `null` and retry, the above code will work: ```python ## In llama_index.embeddings.OllamaEmbedding while True: response = requests.post( url=f"{self.base_url}/api/embeddings", headers={"Content-Type": "application/json"}, json=ollama_request_body, ) response.encoding = "utf-8" if response.status_code != 200: optional_detail = response.json().get("error") raise ValueError( f"Ollama call failed with status code {response.status_code}." f" Details: {optional_detail}" ) if response.json()["embedding"]: break ```
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1707/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1707/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7605
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7605/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7605/comments
https://api.github.com/repos/ollama/ollama/issues/7605/events
https://github.com/ollama/ollama/issues/7605
2,647,941,751
I_kwDOJ0Z1Ps6d1GJ3
7,605
[Model request]GOT-OCR2.0
{ "login": "monkeycc", "id": 6490927, "node_id": "MDQ6VXNlcjY0OTA5Mjc=", "avatar_url": "https://avatars.githubusercontent.com/u/6490927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/monkeycc", "html_url": "https://github.com/monkeycc", "followers_url": "https://api.github.com/users/monkeycc/followers", "following_url": "https://api.github.com/users/monkeycc/following{/other_user}", "gists_url": "https://api.github.com/users/monkeycc/gists{/gist_id}", "starred_url": "https://api.github.com/users/monkeycc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/monkeycc/subscriptions", "organizations_url": "https://api.github.com/users/monkeycc/orgs", "repos_url": "https://api.github.com/users/monkeycc/repos", "events_url": "https://api.github.com/users/monkeycc/events{/privacy}", "received_events_url": "https://api.github.com/users/monkeycc/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2024-11-11T02:30:06
2024-12-18T01:46:27
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://github.com/Ucas-HaoranWei/GOT-OCR2.0
{ "login": "monkeycc", "id": 6490927, "node_id": "MDQ6VXNlcjY0OTA5Mjc=", "avatar_url": "https://avatars.githubusercontent.com/u/6490927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/monkeycc", "html_url": "https://github.com/monkeycc", "followers_url": "https://api.github.com/users/monkeycc/followers", "following_url": "https://api.github.com/users/monkeycc/following{/other_user}", "gists_url": "https://api.github.com/users/monkeycc/gists{/gist_id}", "starred_url": "https://api.github.com/users/monkeycc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/monkeycc/subscriptions", "organizations_url": "https://api.github.com/users/monkeycc/orgs", "repos_url": "https://api.github.com/users/monkeycc/repos", "events_url": "https://api.github.com/users/monkeycc/events{/privacy}", "received_events_url": "https://api.github.com/users/monkeycc/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7605/reactions", "total_count": 6, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7605/timeline
null
reopened
false
https://api.github.com/repos/ollama/ollama/issues/7774
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7774/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7774/comments
https://api.github.com/repos/ollama/ollama/issues/7774/events
https://github.com/ollama/ollama/issues/7774
2,677,843,192
I_kwDOJ0Z1Ps6fnKT4
7,774
Toolcall stream
{ "login": "ZHOUxiaohe1987", "id": 59469405, "node_id": "MDQ6VXNlcjU5NDY5NDA1", "avatar_url": "https://avatars.githubusercontent.com/u/59469405?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ZHOUxiaohe1987", "html_url": "https://github.com/ZHOUxiaohe1987", "followers_url": "https://api.github.com/users/ZHOUxiaohe1987/followers", "following_url": "https://api.github.com/users/ZHOUxiaohe1987/following{/other_user}", "gists_url": "https://api.github.com/users/ZHOUxiaohe1987/gists{/gist_id}", "starred_url": "https://api.github.com/users/ZHOUxiaohe1987/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZHOUxiaohe1987/subscriptions", "organizations_url": "https://api.github.com/users/ZHOUxiaohe1987/orgs", "repos_url": "https://api.github.com/users/ZHOUxiaohe1987/repos", "events_url": "https://api.github.com/users/ZHOUxiaohe1987/events{/privacy}", "received_events_url": "https://api.github.com/users/ZHOUxiaohe1987/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
0
2024-11-21T03:21:31
2024-11-21T03:22:11
2024-11-21T03:21:38
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
null
{ "login": "ZHOUxiaohe1987", "id": 59469405, "node_id": "MDQ6VXNlcjU5NDY5NDA1", "avatar_url": "https://avatars.githubusercontent.com/u/59469405?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ZHOUxiaohe1987", "html_url": "https://github.com/ZHOUxiaohe1987", "followers_url": "https://api.github.com/users/ZHOUxiaohe1987/followers", "following_url": "https://api.github.com/users/ZHOUxiaohe1987/following{/other_user}", "gists_url": "https://api.github.com/users/ZHOUxiaohe1987/gists{/gist_id}", "starred_url": "https://api.github.com/users/ZHOUxiaohe1987/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ZHOUxiaohe1987/subscriptions", "organizations_url": "https://api.github.com/users/ZHOUxiaohe1987/orgs", "repos_url": "https://api.github.com/users/ZHOUxiaohe1987/repos", "events_url": "https://api.github.com/users/ZHOUxiaohe1987/events{/privacy}", "received_events_url": "https://api.github.com/users/ZHOUxiaohe1987/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7774/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7774/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8597
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8597/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8597/comments
https://api.github.com/repos/ollama/ollama/issues/8597/events
https://github.com/ollama/ollama/issues/8597
2,811,701,408
I_kwDOJ0Z1Ps6nlyig
8,597
Error: llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer (4x L40S, 384GB system RAM, Deepseek-R1)
{ "login": "orlyandico", "id": 1325420, "node_id": "MDQ6VXNlcjEzMjU0MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orlyandico", "html_url": "https://github.com/orlyandico", "followers_url": "https://api.github.com/users/orlyandico/followers", "following_url": "https://api.github.com/users/orlyandico/following{/other_user}", "gists_url": "https://api.github.com/users/orlyandico/gists{/gist_id}", "starred_url": "https://api.github.com/users/orlyandico/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/orlyandico/subscriptions", "organizations_url": "https://api.github.com/users/orlyandico/orgs", "repos_url": "https://api.github.com/users/orlyandico/repos", "events_url": "https://api.github.com/users/orlyandico/events{/privacy}", "received_events_url": "https://api.github.com/users/orlyandico/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
17
2025-01-26T17:56:11
2025-01-30T07:47:39
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am aware that 4 x L40S only has 192GB of VRAM, but the machine also has 384GB of system RAM. The error seems to indicate that 45108 MiB of RAM is being allocated with cudaMalloc and this is failing. This is very close to the GPU limit (46068 MiB). On my home setup (2x P40, admittedly not trying such a huge model) the GPU never gets close to its VRAM limit (typically only 22GB out of 24GB). ``` Jan 26 17:48:20 ip-172-31-3-6 ollama[1418]: llm_load_print_meta: rope_yarn_log_mul = 0.1000 Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: ggml_backend_cuda_buffer_type_alloc_buffer: allocating 45108.64 MiB on device 0: cudaMalloc failed: out of memory Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: llama_model_load: error loading model: unable to allocate CUDA0 buffer Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: llama_load_model_from_file: failed to load model Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: panic: unable to load model: /usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: goroutine 34 [running]: Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner.(*Server).loadModel(0xc0001a0000, {0x1b, 0x0, 0x0, 0x0, {0xc000194090, 0x4, 0x4}, 0xc00018a060, 0x0}, ...) Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner/runner.go:852 +0x3ad Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: created by github.com/ollama/ollama/llama/runner.Execute in goroutine 1 Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner/runner.go:970 +0xd0d Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:24.077Z level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server error" Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:24.578Z level=ERROR source=sched.go:455 msg="error loading llama server" error="llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer\nllama_load_model_from_file: failed to load model" Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: [GIN] 2025/01/26 - 17:51:24 | 500 | 3m6s | 127.0.0.1 | POST "/api/generate" Jan 26 17:51:30 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:30.190Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=5.6116838510000004 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1> Jan 26 17:51:30 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:30.947Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=6.369278328 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 26 17:51:31 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:31.707Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=7.128851916 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 26 17:51:40 ip-172-31-3-6 ollama[1418]: [GIN] 2025/01/26 - 17:51:40 | 200 | 65.972µs | 127.0.0.1 | GET "/api/version" ubuntu@ip-172-31-3-6:~$ ollama --version ollama version is 0.5.7 ubuntu@ip-172-31-3-6:~$ free total used free shared buff/cache available Mem: 390837004 4152012 385439624 3248 1245368 383763992 Swap: 0 0 0 ubuntu@ip-172-31-3-6:~$ nvidia-smi Sun Jan 26 17:48:08 2025 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.144.03 Driver Version: 550.144.03 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA L40S On | 00000000:38:00.0 Off | 0 | | N/A 22C P8 22W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA L40S On | 00000000:3A:00.0 Off | 0 | | N/A 22C P8 21W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 2 NVIDIA L40S On | 00000000:3C:00.0 Off | 0 | | N/A 22C P8 44W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 3 NVIDIA L40S On | 00000000:3E:00.0 Off | 0 | | N/A 23C P8 22W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ ``` ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8597/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8597/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7270
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7270/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7270/comments
https://api.github.com/repos/ollama/ollama/issues/7270/events
https://github.com/ollama/ollama/pull/7270
2,599,159,642
PR_kwDOJ0Z1Ps5_LPMN
7,270
Show supported types list in help message for `ollama create`
{ "login": "shirayu", "id": 963961, "node_id": "MDQ6VXNlcjk2Mzk2MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/963961?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shirayu", "html_url": "https://github.com/shirayu", "followers_url": "https://api.github.com/users/shirayu/followers", "following_url": "https://api.github.com/users/shirayu/following{/other_user}", "gists_url": "https://api.github.com/users/shirayu/gists{/gist_id}", "starred_url": "https://api.github.com/users/shirayu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shirayu/subscriptions", "organizations_url": "https://api.github.com/users/shirayu/orgs", "repos_url": "https://api.github.com/users/shirayu/repos", "events_url": "https://api.github.com/users/shirayu/events{/privacy}", "received_events_url": "https://api.github.com/users/shirayu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-10-19T13:15:48
2024-11-27T06:43:00
2024-11-27T06:43:00
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7270", "html_url": "https://github.com/ollama/ollama/pull/7270", "diff_url": "https://github.com/ollama/ollama/pull/7270.diff", "patch_url": "https://github.com/ollama/ollama/pull/7270.patch", "merged_at": null }
This PR adds supported types list to the help message for `ollama create` ```console $ go run main.go create -h Create a model from a Modelfile Usage: ollama create MODEL [flags] Flags: -f, --file string Name of the Modelfile (default "Modelfile") -h, --help help for create -q, --quantize string Quantize model to this level (e.g. q4_0) Supported types: BF16, F16, F32, IQ1_M, IQ1_S, IQ2_M IQ2_S, IQ2_XS, IQ2_XXS, IQ3_S, IQ3_XS, IQ3_XXS IQ4_NL, IQ4_XS, Q2_K, Q2_K_S, Q3_K_L, Q3_K_M Q3_K_S, Q4_0, Q4_1, Q4_1_F16, Q4_K_M, Q4_K_S Q5_0, Q5_1, Q5_K_M, Q5_K_S, Q6_K, Q8_0 Environment Variables: OLLAMA_HOST IP Address for the ollama server (default 127.0.0.1:11434) ```
{ "login": "shirayu", "id": 963961, "node_id": "MDQ6VXNlcjk2Mzk2MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/963961?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shirayu", "html_url": "https://github.com/shirayu", "followers_url": "https://api.github.com/users/shirayu/followers", "following_url": "https://api.github.com/users/shirayu/following{/other_user}", "gists_url": "https://api.github.com/users/shirayu/gists{/gist_id}", "starred_url": "https://api.github.com/users/shirayu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shirayu/subscriptions", "organizations_url": "https://api.github.com/users/shirayu/orgs", "repos_url": "https://api.github.com/users/shirayu/repos", "events_url": "https://api.github.com/users/shirayu/events{/privacy}", "received_events_url": "https://api.github.com/users/shirayu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7270/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7270/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/478
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/478/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/478/comments
https://api.github.com/repos/ollama/ollama/issues/478/events
https://github.com/ollama/ollama/pull/478
1,884,825,856
PR_kwDOJ0Z1Ps5ZufRs
478
remove unused openssh key types
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-09-06T22:13:05
2023-09-06T22:18:55
2023-09-06T22:18:54
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/478", "html_url": "https://github.com/ollama/ollama/pull/478", "diff_url": "https://github.com/ollama/ollama/pull/478.diff", "patch_url": "https://github.com/ollama/ollama/pull/478.patch", "merged_at": "2023-09-06T22:18:54" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/478/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/478/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7369
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7369/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7369/comments
https://api.github.com/repos/ollama/ollama/issues/7369/events
https://github.com/ollama/ollama/pull/7369
2,615,356,549
PR_kwDOJ0Z1Ps5_94Uy
7,369
Fix deepseek deseret regex
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-10-25T23:37:50
2024-10-26T21:58:57
2024-10-26T21:58:54
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7369", "html_url": "https://github.com/ollama/ollama/pull/7369", "diff_url": "https://github.com/ollama/ollama/pull/7369.diff", "patch_url": "https://github.com/ollama/ollama/pull/7369.patch", "merged_at": "2024-10-26T21:58:54" }
On windows compiled with gcc the c++ regex library failed to handle the characters Without any changes, loading the model in the Go server crashes with ``` llama_model_load: error loading model: error loading model vocabulary: wstring_convert::from_bytes ``` The patch for unicode.cpp gets past the wide character conversion problem, but then hits ``` Regex error: Invalid range in bracket expression. ``` Switching to the `\U<8hexchars>` syntax to capture the range for Deseret resolves the regex problem. Fixes #7311
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7369/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7369/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4416
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4416/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4416/comments
https://api.github.com/repos/ollama/ollama/issues/4416/events
https://github.com/ollama/ollama/pull/4416
2,294,175,355
PR_kwDOJ0Z1Ps5vVSj2
4,416
don't abort when an invalid model name is used in /save
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-14T01:42:17
2024-05-14T01:48:29
2024-05-14T01:48:28
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4416", "html_url": "https://github.com/ollama/ollama/pull/4416", "diff_url": "https://github.com/ollama/ollama/pull/4416.diff", "patch_url": "https://github.com/ollama/ollama/pull/4416.patch", "merged_at": "2024-05-14T01:48:28" }
Fixes #3852
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4416/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4416/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2774
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2774/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2774/comments
https://api.github.com/repos/ollama/ollama/issues/2774/events
https://github.com/ollama/ollama/issues/2774
2,155,807,968
I_kwDOJ0Z1Ps6AfwTg
2,774
What is the different between /api/generate and /api/chat?
{ "login": "owenzhao", "id": 2182896, "node_id": "MDQ6VXNlcjIxODI4OTY=", "avatar_url": "https://avatars.githubusercontent.com/u/2182896?v=4", "gravatar_id": "", "url": "https://api.github.com/users/owenzhao", "html_url": "https://github.com/owenzhao", "followers_url": "https://api.github.com/users/owenzhao/followers", "following_url": "https://api.github.com/users/owenzhao/following{/other_user}", "gists_url": "https://api.github.com/users/owenzhao/gists{/gist_id}", "starred_url": "https://api.github.com/users/owenzhao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/owenzhao/subscriptions", "organizations_url": "https://api.github.com/users/owenzhao/orgs", "repos_url": "https://api.github.com/users/owenzhao/repos", "events_url": "https://api.github.com/users/owenzhao/events{/privacy}", "received_events_url": "https://api.github.com/users/owenzhao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
16
2024-02-27T06:53:37
2024-09-09T18:39:13
2024-03-12T06:33:16
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I mean if I give them the same prompt and input, the answers will be the same. Right? Then why they are two different API? Or is chat auto context? I mean when using /api/chat, the answer will automatically include the previous conversation? And the /api/generate only answer for the present?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2774/reactions", "total_count": 14, "+1": 14, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2774/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8232
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8232/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8232/comments
https://api.github.com/repos/ollama/ollama/issues/8232/events
https://github.com/ollama/ollama/pull/8232
2,757,927,665
PR_kwDOJ0Z1Ps6GLO5h
8,232
Remove redundant image_url format check in fromChatRequest
{ "login": "muzzlol", "id": 116030715, "node_id": "U_kgDOBup8-w", "avatar_url": "https://avatars.githubusercontent.com/u/116030715?v=4", "gravatar_id": "", "url": "https://api.github.com/users/muzzlol", "html_url": "https://github.com/muzzlol", "followers_url": "https://api.github.com/users/muzzlol/followers", "following_url": "https://api.github.com/users/muzzlol/following{/other_user}", "gists_url": "https://api.github.com/users/muzzlol/gists{/gist_id}", "starred_url": "https://api.github.com/users/muzzlol/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/muzzlol/subscriptions", "organizations_url": "https://api.github.com/users/muzzlol/orgs", "repos_url": "https://api.github.com/users/muzzlol/repos", "events_url": "https://api.github.com/users/muzzlol/events{/privacy}", "received_events_url": "https://api.github.com/users/muzzlol/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-12-24T15:22:37
2025-01-13T19:32:37
2025-01-13T19:32:37
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8232", "html_url": "https://github.com/ollama/ollama/pull/8232", "diff_url": "https://github.com/ollama/ollama/pull/8232.diff", "patch_url": "https://github.com/ollama/ollama/pull/8232.patch", "merged_at": null }
Improves the `fromChatRequest` function by removing redundant checks for Ollama's direct string format and improving error messages for better debugging. Summary of changes: - Removed the redundant check for `image_url` in Ollama format. - Clarified and enhanced error messages in the conversion logic.
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8232/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8232/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8604
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8604/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8604/comments
https://api.github.com/repos/ollama/ollama/issues/8604/events
https://github.com/ollama/ollama/pull/8604
2,812,382,560
PR_kwDOJ0Z1Ps6JDQsP
8,604
readme: add AI Toolkit for VSCode to community integrations
{ "login": "XiaofuHuang", "id": 49138419, "node_id": "MDQ6VXNlcjQ5MTM4NDE5", "avatar_url": "https://avatars.githubusercontent.com/u/49138419?v=4", "gravatar_id": "", "url": "https://api.github.com/users/XiaofuHuang", "html_url": "https://github.com/XiaofuHuang", "followers_url": "https://api.github.com/users/XiaofuHuang/followers", "following_url": "https://api.github.com/users/XiaofuHuang/following{/other_user}", "gists_url": "https://api.github.com/users/XiaofuHuang/gists{/gist_id}", "starred_url": "https://api.github.com/users/XiaofuHuang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/XiaofuHuang/subscriptions", "organizations_url": "https://api.github.com/users/XiaofuHuang/orgs", "repos_url": "https://api.github.com/users/XiaofuHuang/repos", "events_url": "https://api.github.com/users/XiaofuHuang/events{/privacy}", "received_events_url": "https://api.github.com/users/XiaofuHuang/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2025-01-27T08:34:26
2025-01-30T13:12:11
2025-01-27T08:36:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8604", "html_url": "https://github.com/ollama/ollama/pull/8604", "diff_url": "https://github.com/ollama/ollama/pull/8604.diff", "patch_url": "https://github.com/ollama/ollama/pull/8604.patch", "merged_at": "2025-01-27T08:36:23" }
Add [AI Toolkit for VSCode](https://marketplace.visualstudio.com/items?itemName=ms-windows-ai-studio.windows-ai-studio) to community integrations. AI Toolkit for VSCode ollama support: https://aka.ms/ai-tooklit/ollama-docs
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8604/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8604/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2949
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2949/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2949/comments
https://api.github.com/repos/ollama/ollama/issues/2949/events
https://github.com/ollama/ollama/pull/2949
2,171,044,978
PR_kwDOJ0Z1Ps5o0Jnk
2,949
add api.NewClient
{ "login": "jackielii", "id": 360983, "node_id": "MDQ6VXNlcjM2MDk4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jackielii", "html_url": "https://github.com/jackielii", "followers_url": "https://api.github.com/users/jackielii/followers", "following_url": "https://api.github.com/users/jackielii/following{/other_user}", "gists_url": "https://api.github.com/users/jackielii/gists{/gist_id}", "starred_url": "https://api.github.com/users/jackielii/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jackielii/subscriptions", "organizations_url": "https://api.github.com/users/jackielii/orgs", "repos_url": "https://api.github.com/users/jackielii/repos", "events_url": "https://api.github.com/users/jackielii/events{/privacy}", "received_events_url": "https://api.github.com/users/jackielii/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-03-06T09:31:58
2024-05-04T08:35:08
2024-05-04T08:35:07
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2949", "html_url": "https://github.com/ollama/ollama/pull/2949", "diff_url": "https://github.com/ollama/ollama/pull/2949.diff", "patch_url": "https://github.com/ollama/ollama/pull/2949.patch", "merged_at": null }
fix #2948
{ "login": "jackielii", "id": 360983, "node_id": "MDQ6VXNlcjM2MDk4Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jackielii", "html_url": "https://github.com/jackielii", "followers_url": "https://api.github.com/users/jackielii/followers", "following_url": "https://api.github.com/users/jackielii/following{/other_user}", "gists_url": "https://api.github.com/users/jackielii/gists{/gist_id}", "starred_url": "https://api.github.com/users/jackielii/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jackielii/subscriptions", "organizations_url": "https://api.github.com/users/jackielii/orgs", "repos_url": "https://api.github.com/users/jackielii/repos", "events_url": "https://api.github.com/users/jackielii/events{/privacy}", "received_events_url": "https://api.github.com/users/jackielii/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2949/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2949/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1374
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1374/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1374/comments
https://api.github.com/repos/ollama/ollama/issues/1374/events
https://github.com/ollama/ollama/issues/1374
2,024,279,782
I_kwDOJ0Z1Ps54qA7m
1,374
Out of memory error on model that previously worked fine after update to version 0.1.13
{ "login": "madsamjp", "id": 49611363, "node_id": "MDQ6VXNlcjQ5NjExMzYz", "avatar_url": "https://avatars.githubusercontent.com/u/49611363?v=4", "gravatar_id": "", "url": "https://api.github.com/users/madsamjp", "html_url": "https://github.com/madsamjp", "followers_url": "https://api.github.com/users/madsamjp/followers", "following_url": "https://api.github.com/users/madsamjp/following{/other_user}", "gists_url": "https://api.github.com/users/madsamjp/gists{/gist_id}", "starred_url": "https://api.github.com/users/madsamjp/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/madsamjp/subscriptions", "organizations_url": "https://api.github.com/users/madsamjp/orgs", "repos_url": "https://api.github.com/users/madsamjp/repos", "events_url": "https://api.github.com/users/madsamjp/events{/privacy}", "received_events_url": "https://api.github.com/users/madsamjp/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
27
2023-12-04T16:40:23
2024-01-08T21:42:02
2024-01-08T21:42:02
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I configured a model to run entirely in VRAM using the following Modelfile: ``` FROM deepseek-coder:33b-instruct-q5_K_S PARAMETER num_gpu 65 PARAMETER num_ctx 2048 ``` I had no issues with running this, it would use about 22GB of my 4090's 24GB VRAM without issue. It would generate responses very quickly which was very helpful for getting quick answers to short coding queries. However, yesterday I updated Ollama (to 0.1.13), and now I cannot run the same model. I get an out of memory error, despite the model not needing more than 22.5GB (according to the logs below). I run Ollama on a headless linux server, so there are no other applications using the GPU. Was there an update that changes how much VRAM Ollama allocates to make it need more than before? Is there a way to configure Ollama so that it behaves the same way as before? EDIT: Reverting back to ollama version 0.1.11 resolves the issue for now. Error: ``` Dec 04 16:28:20 osm-server ollama[528776]: llm_load_tensors: offloaded 65/65 layers to GPU Dec 04 16:28:20 osm-server ollama[528776]: llm_load_tensors: VRAM used: 21741.89 MiB Dec 04 16:28:23 osm-server ollama[528776]: .................................................................................................... Dec 04 16:28:23 osm-server ollama[528776]: llama_new_context_with_model: n_ctx = 2048 Dec 04 16:28:23 osm-server ollama[528776]: llama_new_context_with_model: freq_base = 100000.0 Dec 04 16:28:23 osm-server ollama[528776]: llama_new_context_with_model: freq_scale = 0.25 Dec 04 16:28:24 osm-server ollama[528776]: llama_kv_cache_init: offloading v cache to GPU Dec 04 16:28:24 osm-server ollama[528776]: llama_kv_cache_init: offloading k cache to GPU Dec 04 16:28:24 osm-server ollama[528776]: llama_kv_cache_init: VRAM kv self = 496.00 MiB Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: kv self size = 496.00 MiB Dec 04 16:28:24 osm-server ollama[528776]: llama_build_graph: non-view tensors processed: 1430/1430 Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: compute buffer total size = 273.07 MiB Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: VRAM scratch buffer: 270.00 MiB Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: total VRAM used: 22507.89 MiB (model: 21741.89 MiB, context: 766.00 MiB) Dec 04 16:28:24 osm-server ollama[600735]: {"timestamp":1701707304,"level":"INFO","function":"main","line":2917,"message":"HTTP server listening","hostname":"127.0.0.1","port":57264} Dec 04 16:28:24 osm-server ollama[600735]: {"timestamp":1701707304,"level":"INFO","function":"log_server_request","line":2478,"message":"request","remote_addr":"127.0.0.1","remote_port":46990,"status":200,"method":"HEAD","path":"/","params":{}} Dec 04 16:28:24 osm-server ollama[528776]: 2023/12/04 16:28:24 llama.go:493: llama runner started in 4.401485 seconds Dec 04 16:28:24 osm-server ollama[528776]: CUDA error 2 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:5884: out of memory Dec 04 16:28:24 osm-server ollama[528776]: current device: 0 Dec 04 16:28:25 osm-server ollama[528776]: 2023/12/04 16:28:25 llama.go:436: 2 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:5884: out of memory Dec 04 16:28:25 osm-server ollama[528776]: current device: 0 Dec 04 16:28:25 osm-server ollama[528776]: 2023/12/04 16:28:25 llama.go:510: llama runner stopped successfully Dec 04 16:28:25 osm-server ollama[528776]: [GIN] 2023/12/04 - 16:28:25 | 200 | 6.468638351s | 127.0.0.1 | POST "/api/generate" ```
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1374/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1374/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/938
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/938/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/938/comments
https://api.github.com/repos/ollama/ollama/issues/938/events
https://github.com/ollama/ollama/issues/938
1,966,217,268
I_kwDOJ0Z1Ps51Mhg0
938
CLI windows size detection not accurate on WSL
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
0
2023-10-27T21:33:22
2024-02-20T00:56:13
2024-02-20T00:56:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When backspacing characters from a long input on wsl a "float extra character" may not be deleted. Reproduce by going to a newline and backspacing from the last character. <img width="717" alt="Screenshot 2023-10-27 at 17 30 48" src="https://github.com/jmorganca/ollama/assets/5853428/9e18a84b-3071-4b67-aaba-43f2e2d4c1f4"> <img width="716" alt="Screenshot 2023-10-27 at 17 31 12" src="https://github.com/jmorganca/ollama/assets/5853428/029e7b25-d49a-435e-b320-773a22d81369">
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/938/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/938/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3458
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3458/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3458/comments
https://api.github.com/repos/ollama/ollama/issues/3458/events
https://github.com/ollama/ollama/pull/3458
2,220,673,215
PR_kwDOJ0Z1Ps5rcjJg
3,458
fix model needLoad always be true
{ "login": "mofanke", "id": 54242816, "node_id": "MDQ6VXNlcjU0MjQyODE2", "avatar_url": "https://avatars.githubusercontent.com/u/54242816?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mofanke", "html_url": "https://github.com/mofanke", "followers_url": "https://api.github.com/users/mofanke/followers", "following_url": "https://api.github.com/users/mofanke/following{/other_user}", "gists_url": "https://api.github.com/users/mofanke/gists{/gist_id}", "starred_url": "https://api.github.com/users/mofanke/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mofanke/subscriptions", "organizations_url": "https://api.github.com/users/mofanke/orgs", "repos_url": "https://api.github.com/users/mofanke/repos", "events_url": "https://api.github.com/users/mofanke/events{/privacy}", "received_events_url": "https://api.github.com/users/mofanke/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-04-02T14:26:53
2024-04-03T06:05:55
2024-04-03T06:05:54
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3458", "html_url": "https://github.com/ollama/ollama/pull/3458", "diff_url": "https://github.com/ollama/ollama/pull/3458.diff", "patch_url": "https://github.com/ollama/ollama/pull/3458.patch", "merged_at": null }
![image](https://github.com/ollama/ollama/assets/54242816/d82726b0-9988-42f1-bf1f-3b97b476b504) loaded.Options.Runner.NumGPU was change to 33 or other value by `` func New(model string, adapters, projectors []string, opts *api.Options) `` so `!reflect.DeepEqual(loaded.Options.Runner, opts.Runner)` will always be true, because` opts.Runner.NumGPU` which default to -1
{ "login": "mofanke", "id": 54242816, "node_id": "MDQ6VXNlcjU0MjQyODE2", "avatar_url": "https://avatars.githubusercontent.com/u/54242816?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mofanke", "html_url": "https://github.com/mofanke", "followers_url": "https://api.github.com/users/mofanke/followers", "following_url": "https://api.github.com/users/mofanke/following{/other_user}", "gists_url": "https://api.github.com/users/mofanke/gists{/gist_id}", "starred_url": "https://api.github.com/users/mofanke/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mofanke/subscriptions", "organizations_url": "https://api.github.com/users/mofanke/orgs", "repos_url": "https://api.github.com/users/mofanke/repos", "events_url": "https://api.github.com/users/mofanke/events{/privacy}", "received_events_url": "https://api.github.com/users/mofanke/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3458/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3458/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/729
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/729/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/729/comments
https://api.github.com/repos/ollama/ollama/issues/729/events
https://github.com/ollama/ollama/issues/729
1,931,317,404
I_kwDOJ0Z1Ps5zHZCc
729
Unable to pull models behind the proxy
{ "login": "ilyanoskov", "id": 16443383, "node_id": "MDQ6VXNlcjE2NDQzMzgz", "avatar_url": "https://avatars.githubusercontent.com/u/16443383?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ilyanoskov", "html_url": "https://github.com/ilyanoskov", "followers_url": "https://api.github.com/users/ilyanoskov/followers", "following_url": "https://api.github.com/users/ilyanoskov/following{/other_user}", "gists_url": "https://api.github.com/users/ilyanoskov/gists{/gist_id}", "starred_url": "https://api.github.com/users/ilyanoskov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ilyanoskov/subscriptions", "organizations_url": "https://api.github.com/users/ilyanoskov/orgs", "repos_url": "https://api.github.com/users/ilyanoskov/repos", "events_url": "https://api.github.com/users/ilyanoskov/events{/privacy}", "received_events_url": "https://api.github.com/users/ilyanoskov/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
13
2023-10-07T11:38:39
2024-10-01T11:45:58
2023-10-10T16:59:07
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Dear Maintainers, Thank you very much for creating this project! I need to set up ollama on Linux behind a proxy, and when pulling I get an error: ```download.go:166: couldn't download blob: Get "https:///...../ollama/docker/registry/v2/blobs/...": tls: first record does not look like a TLS handshake``` I have tried these methods and they also did not work for me: - https://github.com/jmorganca/ollama/issues/703#issuecomment-1747857562 - https://github.com/jmorganca/ollama/issues/676#issuecomment-1744722380 - https://github.com/jmorganca/ollama/issues/697 Could you please add a way to configure a proxy for Ollama? That will enable a lot of users that must use a proxy. Thank you very much in advance.
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/729/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/729/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4922
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4922/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4922/comments
https://api.github.com/repos/ollama/ollama/issues/4922/events
https://github.com/ollama/ollama/pull/4922
2,341,320,564
PR_kwDOJ0Z1Ps5x2EeC
4,922
cmd: initial functionality for `ollama show <model>`
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-06-08T00:44:51
2024-06-15T03:03:48
2024-06-14T23:18:33
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
true
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4922", "html_url": "https://github.com/ollama/ollama/pull/4922", "diff_url": "https://github.com/ollama/ollama/pull/4922.diff", "patch_url": "https://github.com/ollama/ollama/pull/4922.patch", "merged_at": null }
null
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4922/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4922/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4631
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4631/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4631/comments
https://api.github.com/repos/ollama/ollama/issues/4631/events
https://github.com/ollama/ollama/issues/4631
2,316,815,049
I_kwDOJ0Z1Ps6KF8rJ
4,631
Repeating answers in an instance.
{ "login": "Protostarship", "id": 125519972, "node_id": "U_kgDOB3tIZA", "avatar_url": "https://avatars.githubusercontent.com/u/125519972?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Protostarship", "html_url": "https://github.com/Protostarship", "followers_url": "https://api.github.com/users/Protostarship/followers", "following_url": "https://api.github.com/users/Protostarship/following{/other_user}", "gists_url": "https://api.github.com/users/Protostarship/gists{/gist_id}", "starred_url": "https://api.github.com/users/Protostarship/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Protostarship/subscriptions", "organizations_url": "https://api.github.com/users/Protostarship/orgs", "repos_url": "https://api.github.com/users/Protostarship/repos", "events_url": "https://api.github.com/users/Protostarship/events{/privacy}", "received_events_url": "https://api.github.com/users/Protostarship/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2024-05-25T07:58:33
2024-05-29T22:52:59
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi, i'm using llama2 models, and when i asked the Ai to explain something, it does responds and explain it, but when it reached the end, it instead print out certain line, ``` [INST: <</INST] ``` then repeats another answer regarding the current questions. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version ollama 0.1.38
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4631/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4631/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6263
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6263/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6263/comments
https://api.github.com/repos/ollama/ollama/issues/6263/events
https://github.com/ollama/ollama/issues/6263
2,456,652,905
I_kwDOJ0Z1Ps6SbYxp
6,263
Pull Command Parsing Not Working
{ "login": "chadwickhar08", "id": 23239905, "node_id": "MDQ6VXNlcjIzMjM5OTA1", "avatar_url": "https://avatars.githubusercontent.com/u/23239905?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chadwickhar08", "html_url": "https://github.com/chadwickhar08", "followers_url": "https://api.github.com/users/chadwickhar08/followers", "following_url": "https://api.github.com/users/chadwickhar08/following{/other_user}", "gists_url": "https://api.github.com/users/chadwickhar08/gists{/gist_id}", "starred_url": "https://api.github.com/users/chadwickhar08/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chadwickhar08/subscriptions", "organizations_url": "https://api.github.com/users/chadwickhar08/orgs", "repos_url": "https://api.github.com/users/chadwickhar08/repos", "events_url": "https://api.github.com/users/chadwickhar08/events{/privacy}", "received_events_url": "https://api.github.com/users/chadwickhar08/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
9
2024-08-08T21:27:30
2024-08-10T00:18:27
2024-08-09T19:30:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When running ollama on Windows, attempt to run 'ollama pull llama3.1' results in 'ollama pull llama3.1 pulling manifest Error: Incorrect function.' ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version ollama --version ollama version is 0.3.4
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6263/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6263/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6412
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6412/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6412/comments
https://api.github.com/repos/ollama/ollama/issues/6412/events
https://github.com/ollama/ollama/issues/6412
2,472,517,560
I_kwDOJ0Z1Ps6TX5-4
6,412
for glm4-9b
{ "login": "lcck200121", "id": 126126214, "node_id": "U_kgDOB4SIhg", "avatar_url": "https://avatars.githubusercontent.com/u/126126214?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lcck200121", "html_url": "https://github.com/lcck200121", "followers_url": "https://api.github.com/users/lcck200121/followers", "following_url": "https://api.github.com/users/lcck200121/following{/other_user}", "gists_url": "https://api.github.com/users/lcck200121/gists{/gist_id}", "starred_url": "https://api.github.com/users/lcck200121/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lcck200121/subscriptions", "organizations_url": "https://api.github.com/users/lcck200121/orgs", "repos_url": "https://api.github.com/users/lcck200121/repos", "events_url": "https://api.github.com/users/lcck200121/events{/privacy}", "received_events_url": "https://api.github.com/users/lcck200121/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2024-08-19T05:54:49
2024-08-19T05:54:49
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Do you have any plans to support the tool-calling function of glm4-9b?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6412/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6412/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4194
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4194/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4194/comments
https://api.github.com/repos/ollama/ollama/issues/4194/events
https://github.com/ollama/ollama/pull/4194
2,280,062,512
PR_kwDOJ0Z1Ps5umAi_
4,194
docs: delete log `HEAD`
{ "login": "HydenLiu", "id": 52030677, "node_id": "MDQ6VXNlcjUyMDMwNjc3", "avatar_url": "https://avatars.githubusercontent.com/u/52030677?v=4", "gravatar_id": "", "url": "https://api.github.com/users/HydenLiu", "html_url": "https://github.com/HydenLiu", "followers_url": "https://api.github.com/users/HydenLiu/followers", "following_url": "https://api.github.com/users/HydenLiu/following{/other_user}", "gists_url": "https://api.github.com/users/HydenLiu/gists{/gist_id}", "starred_url": "https://api.github.com/users/HydenLiu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/HydenLiu/subscriptions", "organizations_url": "https://api.github.com/users/HydenLiu/orgs", "repos_url": "https://api.github.com/users/HydenLiu/repos", "events_url": "https://api.github.com/users/HydenLiu/events{/privacy}", "received_events_url": "https://api.github.com/users/HydenLiu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-06T05:28:47
2024-05-06T17:32:31
2024-05-06T17:32:31
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4194", "html_url": "https://github.com/ollama/ollama/pull/4194", "diff_url": "https://github.com/ollama/ollama/pull/4194.diff", "patch_url": "https://github.com/ollama/ollama/pull/4194.patch", "merged_at": "2024-05-06T17:32:31" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4194/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4194/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1804
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1804/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1804/comments
https://api.github.com/repos/ollama/ollama/issues/1804/events
https://github.com/ollama/ollama/issues/1804
2,067,249,992
I_kwDOJ0Z1Ps57N7tI
1,804
Code view on codellama vs phi and dolphin-phi
{ "login": "oliverbob", "id": 23272429, "node_id": "MDQ6VXNlcjIzMjcyNDI5", "avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/oliverbob", "html_url": "https://github.com/oliverbob", "followers_url": "https://api.github.com/users/oliverbob/followers", "following_url": "https://api.github.com/users/oliverbob/following{/other_user}", "gists_url": "https://api.github.com/users/oliverbob/gists{/gist_id}", "starred_url": "https://api.github.com/users/oliverbob/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/oliverbob/subscriptions", "organizations_url": "https://api.github.com/users/oliverbob/orgs", "repos_url": "https://api.github.com/users/oliverbob/repos", "events_url": "https://api.github.com/users/oliverbob/events{/privacy}", "received_events_url": "https://api.github.com/users/oliverbob/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-05T12:19:46
2024-03-13T23:53:54
2024-03-13T23:53:54
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Dolphin phi and (probably phi code indent): ![image](https://github.com/jmorganca/ollama/assets/23272429/6efbf418-0cbd-46bf-abf3-005db9e2fc3d) ![image](https://github.com/jmorganca/ollama/assets/23272429/5d0a658f-ffd8-44f2-b3f5-992b722d3c37) Phi, indents but has no code view: ![image](https://github.com/jmorganca/ollama/assets/23272429/9cefc3b5-5bf5-4620-af2b-a9259c036c94) Can someone probably do something to improve these models as they are the only models that run very fast on smaller GPUs. Or perhaps, maybe someone would train phi-code:instruct. Thanks.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1804/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1804/timeline
null
completed
false