url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/3855
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3855/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3855/comments
https://api.github.com/repos/ollama/ollama/issues/3855/events
https://github.com/ollama/ollama/pull/3855
2,259,852,585
PR_kwDOJ0Z1Ps5thxhS
3,855
Update windows.md
{ "login": "fyxtro", "id": 16174614, "node_id": "MDQ6VXNlcjE2MTc0NjE0", "avatar_url": "https://avatars.githubusercontent.com/u/16174614?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fyxtro", "html_url": "https://github.com/fyxtro", "followers_url": "https://api.github.com/users/fyxtro/followers", "following_url": "https://api.github.com/users/fyxtro/following{/other_user}", "gists_url": "https://api.github.com/users/fyxtro/gists{/gist_id}", "starred_url": "https://api.github.com/users/fyxtro/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fyxtro/subscriptions", "organizations_url": "https://api.github.com/users/fyxtro/orgs", "repos_url": "https://api.github.com/users/fyxtro/repos", "events_url": "https://api.github.com/users/fyxtro/events{/privacy}", "received_events_url": "https://api.github.com/users/fyxtro/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-23T22:02:07
2024-04-26T20:04:15
2024-04-26T20:04:15
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3855", "html_url": "https://github.com/ollama/ollama/pull/3855", "diff_url": "https://github.com/ollama/ollama/pull/3855.diff", "patch_url": "https://github.com/ollama/ollama/pull/3855.patch", "merged_at": "2024-04-26T20:04:15" }
Fixed a typo
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3855/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3855/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5206
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5206/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5206/comments
https://api.github.com/repos/ollama/ollama/issues/5206/events
https://github.com/ollama/ollama/pull/5206
2,367,206,966
PR_kwDOJ0Z1Ps5zOINU
5,206
fix: quantization with template
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-21T20:30:52
2024-06-21T20:44:35
2024-06-21T20:44:35
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5206", "html_url": "https://github.com/ollama/ollama/pull/5206", "diff_url": "https://github.com/ollama/ollama/pull/5206.diff", "patch_url": "https://github.com/ollama/ollama/pull/5206.patch", "merged_at": "2024-06-21T20:44:35" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5206/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5206/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1975
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1975/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1975/comments
https://api.github.com/repos/ollama/ollama/issues/1975/events
https://github.com/ollama/ollama/issues/1975
2,080,218,641
I_kwDOJ0Z1Ps57_Z4R
1,975
Will ollama run dolphin-mixtral on my gtx 1080 Ti?
{ "login": "PixelovyLabyrintDev", "id": 146265327, "node_id": "U_kgDOCLfU7w", "avatar_url": "https://avatars.githubusercontent.com/u/146265327?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PixelovyLabyrintDev", "html_url": "https://github.com/PixelovyLabyrintDev", "followers_url": "https://api.github.com/users/PixelovyLabyrintDev/followers", "following_url": "https://api.github.com/users/PixelovyLabyrintDev/following{/other_user}", "gists_url": "https://api.github.com/users/PixelovyLabyrintDev/gists{/gist_id}", "starred_url": "https://api.github.com/users/PixelovyLabyrintDev/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/PixelovyLabyrintDev/subscriptions", "organizations_url": "https://api.github.com/users/PixelovyLabyrintDev/orgs", "repos_url": "https://api.github.com/users/PixelovyLabyrintDev/repos", "events_url": "https://api.github.com/users/PixelovyLabyrintDev/events{/privacy}", "received_events_url": "https://api.github.com/users/PixelovyLabyrintDev/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
3
2024-01-13T10:07:52
2024-01-15T21:02:32
2024-01-15T21:02:27
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Im just asking since im about to buy one and im curious if it will see the gpu and use it to generate responses faster? Or does Ollama support all Nvidia gpus?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1975/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1975/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6444
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6444/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6444/comments
https://api.github.com/repos/ollama/ollama/issues/6444/events
https://github.com/ollama/ollama/issues/6444
2,475,946,220
I_kwDOJ0Z1Ps6Tk_Ds
6,444
Update model parameters for SmolLM (and other models)
{ "login": "DuckyBlender", "id": 42645784, "node_id": "MDQ6VXNlcjQyNjQ1Nzg0", "avatar_url": "https://avatars.githubusercontent.com/u/42645784?v=4", "gravatar_id": "", "url": "https://api.github.com/users/DuckyBlender", "html_url": "https://github.com/DuckyBlender", "followers_url": "https://api.github.com/users/DuckyBlender/followers", "following_url": "https://api.github.com/users/DuckyBlender/following{/other_user}", "gists_url": "https://api.github.com/users/DuckyBlender/gists{/gist_id}", "starred_url": "https://api.github.com/users/DuckyBlender/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/DuckyBlender/subscriptions", "organizations_url": "https://api.github.com/users/DuckyBlender/orgs", "repos_url": "https://api.github.com/users/DuckyBlender/repos", "events_url": "https://api.github.com/users/DuckyBlender/events{/privacy}", "received_events_url": "https://api.github.com/users/DuckyBlender/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-08-20T15:31:02
2024-08-21T19:56:25
2024-08-21T19:56:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? It is crucial to follow the recommended model parameters, especially for such small models such as SmolLM. They are much worse at higher temperatures. Ollama team, respectfully, PLEASE make sure the model parameters are set correctly 🙏 From the SmolLM HF: ![firefox_n1LVtmDwfP](https://github.com/user-attachments/assets/967375c9-eefa-4a12-a0d9-d3cdc3386a94) ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6444/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6444/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5684
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5684/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5684/comments
https://api.github.com/repos/ollama/ollama/issues/5684/events
https://github.com/ollama/ollama/pull/5684
2,407,219,014
PR_kwDOJ0Z1Ps51TsEO
5,684
add chat and generate tests with mock runner
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-14T01:07:47
2024-07-16T16:44:48
2024-07-16T16:44:46
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5684", "html_url": "https://github.com/ollama/ollama/pull/5684", "diff_url": "https://github.com/ollama/ollama/pull/5684.diff", "patch_url": "https://github.com/ollama/ollama/pull/5684.patch", "merged_at": "2024-07-16T16:44:46" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5684/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5684/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8209
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8209/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8209/comments
https://api.github.com/repos/ollama/ollama/issues/8209/events
https://github.com/ollama/ollama/pull/8209
2,754,502,642
PR_kwDOJ0Z1Ps6GAaQz
8,209
api: enable passing images in `raw` mode
{ "login": "rylativity", "id": 41017744, "node_id": "MDQ6VXNlcjQxMDE3NzQ0", "avatar_url": "https://avatars.githubusercontent.com/u/41017744?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rylativity", "html_url": "https://github.com/rylativity", "followers_url": "https://api.github.com/users/rylativity/followers", "following_url": "https://api.github.com/users/rylativity/following{/other_user}", "gists_url": "https://api.github.com/users/rylativity/gists{/gist_id}", "starred_url": "https://api.github.com/users/rylativity/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rylativity/subscriptions", "organizations_url": "https://api.github.com/users/rylativity/orgs", "repos_url": "https://api.github.com/users/rylativity/repos", "events_url": "https://api.github.com/users/rylativity/events{/privacy}", "received_events_url": "https://api.github.com/users/rylativity/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2024-12-22T05:52:44
2024-12-22T15:22:20
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8209", "html_url": "https://github.com/ollama/ollama/pull/8209", "diff_url": "https://github.com/ollama/ollama/pull/8209.diff", "patch_url": "https://github.com/ollama/ollama/pull/8209.patch", "merged_at": null }
I belive this should be considered a bug fix, as it resolves unexpected behavior and clarifies how to use undocumented capabilities - Images were not being passed to multimodal models when `images` were provided and `raw` was set to true as described in [this issue](https://github.com/ollama/ollama-python/issues/319). This change introduces an optional boolean parameter `prepend_images_to_raw_prompt`, which if set to `true` will add the image placeholder fields necessary to pass the images through to the multimodal model. This parameter only has effect if `raw` is also set to true and `images` are provided. The `prepend_images_to_raw_prompt` parameter can be omitted, and the default behavior remains entirely unchanged to avoid breaking existing users. It also updates the api.md documentation to describe the new functionality (including an example request/response), and also provides and example of manually passing an image placeholder in the text of a raw prompt directly, which offers more control over where the image is inserted in the raw prompt.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8209/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8209/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3666
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3666/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3666/comments
https://api.github.com/repos/ollama/ollama/issues/3666/events
https://github.com/ollama/ollama/issues/3666
2,245,020,495
I_kwDOJ0Z1Ps6F0EtP
3,666
Ollama's API Confusion
{ "login": "17Reset", "id": 122418720, "node_id": "U_kgDOB0v2IA", "avatar_url": "https://avatars.githubusercontent.com/u/122418720?v=4", "gravatar_id": "", "url": "https://api.github.com/users/17Reset", "html_url": "https://github.com/17Reset", "followers_url": "https://api.github.com/users/17Reset/followers", "following_url": "https://api.github.com/users/17Reset/following{/other_user}", "gists_url": "https://api.github.com/users/17Reset/gists{/gist_id}", "starred_url": "https://api.github.com/users/17Reset/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/17Reset/subscriptions", "organizations_url": "https://api.github.com/users/17Reset/orgs", "repos_url": "https://api.github.com/users/17Reset/repos", "events_url": "https://api.github.com/users/17Reset/events{/privacy}", "received_events_url": "https://api.github.com/users/17Reset/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-04-16T03:32:15
2024-04-17T03:04:22
2024-04-16T23:14:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I started two ports in my server as ollama's api service ports, 8080 and 8081, 8080 as my LLM-api, 8081 as my Embedding-api, but when I add a model to 8080, I can actually see the model on port 8080 by using the following command, but the following command It should see the model for port 8081: curl http://192.168.18.165:8081/api/tags
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3666/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3666/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1405
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1405/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1405/comments
https://api.github.com/repos/ollama/ollama/issues/1405/events
https://github.com/ollama/ollama/issues/1405
2,029,284,255
I_kwDOJ0Z1Ps549Guf
1,405
`ollama run` fails on start
{ "login": "j-sneh", "id": 25213198, "node_id": "MDQ6VXNlcjI1MjEzMTk4", "avatar_url": "https://avatars.githubusercontent.com/u/25213198?v=4", "gravatar_id": "", "url": "https://api.github.com/users/j-sneh", "html_url": "https://github.com/j-sneh", "followers_url": "https://api.github.com/users/j-sneh/followers", "following_url": "https://api.github.com/users/j-sneh/following{/other_user}", "gists_url": "https://api.github.com/users/j-sneh/gists{/gist_id}", "starred_url": "https://api.github.com/users/j-sneh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/j-sneh/subscriptions", "organizations_url": "https://api.github.com/users/j-sneh/orgs", "repos_url": "https://api.github.com/users/j-sneh/repos", "events_url": "https://api.github.com/users/j-sneh/events{/privacy}", "received_events_url": "https://api.github.com/users/j-sneh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2023-12-06T19:56:14
2023-12-06T20:40:56
2023-12-06T20:40:56
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
``` ➜ ~ ollama run llama2 ⠋ Error: template: :2:11: executing "" at <.Context>: can't evaluate field Context in type struct { First bool; System string; Prompt string } ➜ ~ ollama --version ollama version 0.1.13 ``` Ran into this issue after running `ollama run llama2`. Occurs if server is running or if not & issue persisted after updating the server.
{ "login": "j-sneh", "id": 25213198, "node_id": "MDQ6VXNlcjI1MjEzMTk4", "avatar_url": "https://avatars.githubusercontent.com/u/25213198?v=4", "gravatar_id": "", "url": "https://api.github.com/users/j-sneh", "html_url": "https://github.com/j-sneh", "followers_url": "https://api.github.com/users/j-sneh/followers", "following_url": "https://api.github.com/users/j-sneh/following{/other_user}", "gists_url": "https://api.github.com/users/j-sneh/gists{/gist_id}", "starred_url": "https://api.github.com/users/j-sneh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/j-sneh/subscriptions", "organizations_url": "https://api.github.com/users/j-sneh/orgs", "repos_url": "https://api.github.com/users/j-sneh/repos", "events_url": "https://api.github.com/users/j-sneh/events{/privacy}", "received_events_url": "https://api.github.com/users/j-sneh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1405/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1405/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3792
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3792/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3792/comments
https://api.github.com/repos/ollama/ollama/issues/3792/events
https://github.com/ollama/ollama/pull/3792
2,254,927,442
PR_kwDOJ0Z1Ps5tRFfO
3,792
docs: added Ollama Operator into README.md as one of community projects
{ "login": "nekomeowww", "id": 11081491, "node_id": "MDQ6VXNlcjExMDgxNDkx", "avatar_url": "https://avatars.githubusercontent.com/u/11081491?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nekomeowww", "html_url": "https://github.com/nekomeowww", "followers_url": "https://api.github.com/users/nekomeowww/followers", "following_url": "https://api.github.com/users/nekomeowww/following{/other_user}", "gists_url": "https://api.github.com/users/nekomeowww/gists{/gist_id}", "starred_url": "https://api.github.com/users/nekomeowww/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nekomeowww/subscriptions", "organizations_url": "https://api.github.com/users/nekomeowww/orgs", "repos_url": "https://api.github.com/users/nekomeowww/repos", "events_url": "https://api.github.com/users/nekomeowww/events{/privacy}", "received_events_url": "https://api.github.com/users/nekomeowww/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-04-21T06:34:23
2024-11-20T16:33:37
2024-11-20T16:33:37
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3792", "html_url": "https://github.com/ollama/ollama/pull/3792", "diff_url": "https://github.com/ollama/ollama/pull/3792.diff", "patch_url": "https://github.com/ollama/ollama/pull/3792.patch", "merged_at": null }
First of all, huge thank you to all the wonderful work and awesome contributors to you folks for both ollama, llama.cpp and researchers worked on it. It made it easy for us to deploy and host our own large language model. ## Summary I'm Neko Ayaka (https://github.com/nekomeowww) from China. I'm currently working as a full-stack senior developer at @DaoCloud, diving deep into cloud native tech, AI, and UI/UX design. Ollama is decent to deploy as single instance on machines like Macbook, Mac Studio. I'm really pumped about how we designed the ollama cli. It keeps things simple for users, much like what Docker does (it reminds me Docker CLI every time I use it). And the implementation of `Modelfile` in the project really streamlines the process, again reminds me `Dockerfile`. Additionally, I researched the users needs, and interested in over the official Discord server of Ollama, found that many users are researching to find a way to deploy it concurrently. Therefore, inspired by the awesome user experience with ollama, I wanted to bring that same vibe to my own Kubernetes setup at home. That's where the idea for this project started, fueled by conversations with friends who are also into cloud native projects. This is the open-source project called Ollama Operator (GitHub: https://github.com/nekomeowww/ollama-operator, documentation site: https://ollama-operator.ayaka.io/pages/en/ ) I want to introduce and add to README of Ollama here. It's built around a Kubernetes operator concept by leveraging both `ollama pull` and `ollama serve` to make it possible to deploy multiple instances of `ollama serve` to inference multiple models across nodes of clusters by introducing [`Model` CRD](https://ollama-operator.ayaka.io/pages/en/references/crd/model.html), like this: ```yaml apiVersion: ollama.ayaka.io/v1 kind: Model metadata: name: phi spec: # Scale the model to 2 replicas replicas: 2 # Use the model image `phi` image: phi ``` I've mapped out the specs and CRDs for deploying ollama instances on Kubernetes. There are still a few kinks to iron out, but it's looking good as a proof of concept. I've put together all the docs, the architectural design, and even got it up on a neat documentation site powered by VitePress! Besides Ollama Operator, and the introduced `Model` CRD to simplify the process of multi-instance deployment, I've also made a CLI tool called `kollama` (source at: https://github.com/nekomeowww/ollama-operator/tree/main/cmd/kollama , documentation site: https://ollama-operator.ayaka.io/pages/en/references/cli/commands/deploy.html ) to even simplify the interactions with introduced `Model` CRD with just a single command, like this: ```shell kollama deploy phi --expose ``` I have Ollama Operator running on our team's dedicated server, my own development `kind` cluster on my Macbook, and the general K8s cluster on my Homelab, little testing k3s cluster on my two Respberry Pi. They are running smoothly for fine for the past 10 days, and I consider it is a general available project for users to try out and feedback to me to improve the project continously. > There are still much work to do and many things worth me to research and test, I really want to share the concepts, design, simplified concepts with Kubernets. ## Proposal Add and include [Ollama Operator](https://github.com/nekomeowww/ollama-operator) into Ollama's `README.md` file as one of the community driven project.
{ "login": "nekomeowww", "id": 11081491, "node_id": "MDQ6VXNlcjExMDgxNDkx", "avatar_url": "https://avatars.githubusercontent.com/u/11081491?v=4", "gravatar_id": "", "url": "https://api.github.com/users/nekomeowww", "html_url": "https://github.com/nekomeowww", "followers_url": "https://api.github.com/users/nekomeowww/followers", "following_url": "https://api.github.com/users/nekomeowww/following{/other_user}", "gists_url": "https://api.github.com/users/nekomeowww/gists{/gist_id}", "starred_url": "https://api.github.com/users/nekomeowww/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nekomeowww/subscriptions", "organizations_url": "https://api.github.com/users/nekomeowww/orgs", "repos_url": "https://api.github.com/users/nekomeowww/repos", "events_url": "https://api.github.com/users/nekomeowww/events{/privacy}", "received_events_url": "https://api.github.com/users/nekomeowww/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3792/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3792/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7823
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7823/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7823/comments
https://api.github.com/repos/ollama/ollama/issues/7823/events
https://github.com/ollama/ollama/issues/7823
2,688,378,590
I_kwDOJ0Z1Ps6gPWbe
7,823
Multiple prompt support over stdin.
{ "login": "WyvernDotRed", "id": 41121402, "node_id": "MDQ6VXNlcjQxMTIxNDAy", "avatar_url": "https://avatars.githubusercontent.com/u/41121402?v=4", "gravatar_id": "", "url": "https://api.github.com/users/WyvernDotRed", "html_url": "https://github.com/WyvernDotRed", "followers_url": "https://api.github.com/users/WyvernDotRed/followers", "following_url": "https://api.github.com/users/WyvernDotRed/following{/other_user}", "gists_url": "https://api.github.com/users/WyvernDotRed/gists{/gist_id}", "starred_url": "https://api.github.com/users/WyvernDotRed/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/WyvernDotRed/subscriptions", "organizations_url": "https://api.github.com/users/WyvernDotRed/orgs", "repos_url": "https://api.github.com/users/WyvernDotRed/repos", "events_url": "https://api.github.com/users/WyvernDotRed/events{/privacy}", "received_events_url": "https://api.github.com/users/WyvernDotRed/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
5
2024-11-24T20:56:40
2024-12-29T20:26:06
2024-12-29T20:26:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
edit: Important clarifications on a misunderstanding from me on the current implementation at https://github.com/ollama/ollama/issues/7823#issuecomment-2542541753 These points still stay relevant, but might be easier to understand by reading the comment first. In [pull 416](https://github.com/ollama/ollama/pull/416), the ability to enter multiple prompts over STDIN directly seems to have been removed. To work around this, I recently made an `except` script (assembler in Fish) which filters out the interface related outputs elements of `ollama run`. In version 0.4.3, this achieves the desired effect of only outputting model output through the pipe, while allowing for multiple prompts. I will share my script in the first comment. As I mentioned in [issue 7820](https://github.com/ollama/ollama/issues/7820), this method of inputting has been broken as per version 0.4.4. And as mentioned by @rick-github in [their comment](https://github.com/ollama/ollama/issues/7820#issuecomment-2496152849), this likely is by design, due to [pull 7360](https://github.com/ollama/ollama/pull/7360). Which at face value seems to completely cover my use-case and avoids the jank of a script parsing the interface information. But the current implementation breaks entering prompts manually or through `expect` (my [issue 7820](https://github.com/ollama/ollama/issues/7820) again), while piping the output to `tee` or similar. I will explore the proposed `script` based workaround [mentioned by @rick-github](https://github.com/ollama/ollama/issues/7820#issuecomment-2496152849), either adding it to or replacing my `expect` based workaround with it. Like my script, this still is an excessive level of workarounds to make something which would be expected functionality work. **Potential solutions:** - Not having `ollama run` immediately close with output redirection, as the implementation from [pull 7360](https://github.com/ollama/ollama/pull/7360) does. - Having some form of delimiter within the """ syntax, for starting a new prompt, like: ``` """ First prompt. --- Second prompt. """ ``` This could start processing the first/earlier prompt after the --- delimiter is received to allow for full dialogue - Some feature flag to interpret STDIN as user input, similar to [pull 6130](https://github.com/ollama/ollama/pull/6130) (which [pull 7360](https://github.com/ollama/ollama/pull/7360) replaced). Perhaps `--input user`, `--nogui`, `--keeppipe`, ect. - Having multiple prompts passed as arguments processed sequentially, like the shared script does. Of course, this is only a feature *request*, it's fine if piping multiple prompts over STDIN or as arguments is not supported. In which case I would be fine with looking further and interested in suggestions for other frontends or libraries to use. P.S. Since making such a feature request is a first for me, constructive criticism is welcome if you can spare the time. Either way, have a nice rest of your day!
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7823/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7823/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1801
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1801/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1801/comments
https://api.github.com/repos/ollama/ollama/issues/1801/events
https://github.com/ollama/ollama/pull/1801
2,066,643,154
PR_kwDOJ0Z1Ps5jRorQ
1,801
fix docker doc to point to hub
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-05T02:43:54
2024-01-05T03:20:45
2024-01-05T03:20:45
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1801", "html_url": "https://github.com/ollama/ollama/pull/1801", "diff_url": "https://github.com/ollama/ollama/pull/1801.diff", "patch_url": "https://github.com/ollama/ollama/pull/1801.patch", "merged_at": "2024-01-05T03:20:45" }
there is a link that should point to dockerhub which is the best place to understand docker
{ "login": "technovangelist", "id": 633681, "node_id": "MDQ6VXNlcjYzMzY4MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4", "gravatar_id": "", "url": "https://api.github.com/users/technovangelist", "html_url": "https://github.com/technovangelist", "followers_url": "https://api.github.com/users/technovangelist/followers", "following_url": "https://api.github.com/users/technovangelist/following{/other_user}", "gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}", "starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions", "organizations_url": "https://api.github.com/users/technovangelist/orgs", "repos_url": "https://api.github.com/users/technovangelist/repos", "events_url": "https://api.github.com/users/technovangelist/events{/privacy}", "received_events_url": "https://api.github.com/users/technovangelist/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1801/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1801/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7227
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7227/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7227/comments
https://api.github.com/repos/ollama/ollama/issues/7227/events
https://github.com/ollama/ollama/issues/7227
2,592,456,795
I_kwDOJ0Z1Ps6ahcBb
7,227
Ministral-8B-Instruct-2410
{ "login": "vYLQs6", "id": 143073604, "node_id": "U_kgDOCIchRA", "avatar_url": "https://avatars.githubusercontent.com/u/143073604?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vYLQs6", "html_url": "https://github.com/vYLQs6", "followers_url": "https://api.github.com/users/vYLQs6/followers", "following_url": "https://api.github.com/users/vYLQs6/following{/other_user}", "gists_url": "https://api.github.com/users/vYLQs6/gists{/gist_id}", "starred_url": "https://api.github.com/users/vYLQs6/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vYLQs6/subscriptions", "organizations_url": "https://api.github.com/users/vYLQs6/orgs", "repos_url": "https://api.github.com/users/vYLQs6/repos", "events_url": "https://api.github.com/users/vYLQs6/events{/privacy}", "received_events_url": "https://api.github.com/users/vYLQs6/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
4
2024-10-16T16:42:21
2024-10-25T09:15:34
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://huggingface.co/mistralai/Ministral-8B-Instruct-2410 <img width="1633" alt="instruct_table_with_gemma" src="https://github.com/user-attachments/assets/23eb1742-ee5e-475a-bf7b-5853f781c4f7">
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7227/reactions", "total_count": 12, "+1": 11, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/7227/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7441
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7441/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7441/comments
https://api.github.com/repos/ollama/ollama/issues/7441/events
https://github.com/ollama/ollama/issues/7441
2,626,061,571
I_kwDOJ0Z1Ps6choUD
7,441
Error: unknown error was encountered while running the model GGML_ASSERT(i01 >= 0 && i01 < ne01) failed
{ "login": "kalcao", "id": 174179425, "node_id": "U_kgDOCmHEYQ", "avatar_url": "https://avatars.githubusercontent.com/u/174179425?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kalcao", "html_url": "https://github.com/kalcao", "followers_url": "https://api.github.com/users/kalcao/followers", "following_url": "https://api.github.com/users/kalcao/following{/other_user}", "gists_url": "https://api.github.com/users/kalcao/gists{/gist_id}", "starred_url": "https://api.github.com/users/kalcao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kalcao/subscriptions", "organizations_url": "https://api.github.com/users/kalcao/orgs", "repos_url": "https://api.github.com/users/kalcao/repos", "events_url": "https://api.github.com/users/kalcao/events{/privacy}", "received_events_url": "https://api.github.com/users/kalcao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
23
2024-10-31T06:23:20
2025-01-24T16:04:18
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? [Nanollava](https://ollama.com/qnguyen3/nanollava) returns `GGML_ASSERT(i01 >= 0 && i01 < ne01) failed` error on chat with image Output: ```bash ubuntu@ubuntu:~/workspace$ ollama run qnguyen3/nanollava "tell me what do you see in this picture? ./sample.jpg" Added image './sample.jpg' Error: an unknown error was encountered while running the model GGML_ASSERT(i01 >= 0 && i01 < ne01) failed ``` ### OS Linux ### GPU Other ### CPU AMD ### Ollama version 0.3.14
{ "login": "kalcao", "id": 174179425, "node_id": "U_kgDOCmHEYQ", "avatar_url": "https://avatars.githubusercontent.com/u/174179425?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kalcao", "html_url": "https://github.com/kalcao", "followers_url": "https://api.github.com/users/kalcao/followers", "following_url": "https://api.github.com/users/kalcao/following{/other_user}", "gists_url": "https://api.github.com/users/kalcao/gists{/gist_id}", "starred_url": "https://api.github.com/users/kalcao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kalcao/subscriptions", "organizations_url": "https://api.github.com/users/kalcao/orgs", "repos_url": "https://api.github.com/users/kalcao/repos", "events_url": "https://api.github.com/users/kalcao/events{/privacy}", "received_events_url": "https://api.github.com/users/kalcao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7441/reactions", "total_count": 6, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 2 }
https://api.github.com/repos/ollama/ollama/issues/7441/timeline
null
reopened
false
https://api.github.com/repos/ollama/ollama/issues/4109
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4109/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4109/comments
https://api.github.com/repos/ollama/ollama/issues/4109/events
https://github.com/ollama/ollama/pull/4109
2,276,652,765
PR_kwDOJ0Z1Ps5ua2-7
4,109
Add information on model tagging to `import.md`.
{ "login": "thinkverse", "id": 2221746, "node_id": "MDQ6VXNlcjIyMjE3NDY=", "avatar_url": "https://avatars.githubusercontent.com/u/2221746?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thinkverse", "html_url": "https://github.com/thinkverse", "followers_url": "https://api.github.com/users/thinkverse/followers", "following_url": "https://api.github.com/users/thinkverse/following{/other_user}", "gists_url": "https://api.github.com/users/thinkverse/gists{/gist_id}", "starred_url": "https://api.github.com/users/thinkverse/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thinkverse/subscriptions", "organizations_url": "https://api.github.com/users/thinkverse/orgs", "repos_url": "https://api.github.com/users/thinkverse/repos", "events_url": "https://api.github.com/users/thinkverse/events{/privacy}", "received_events_url": "https://api.github.com/users/thinkverse/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-05-02T23:17:24
2024-06-08T23:32:23
2024-06-08T23:32:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4109", "html_url": "https://github.com/ollama/ollama/pull/4109", "diff_url": "https://github.com/ollama/ollama/pull/4109.diff", "patch_url": "https://github.com/ollama/ollama/pull/4109.patch", "merged_at": null }
I added a section to the import documentation on model tagging and cleaned up the quantization reference section a smidge.
{ "login": "thinkverse", "id": 2221746, "node_id": "MDQ6VXNlcjIyMjE3NDY=", "avatar_url": "https://avatars.githubusercontent.com/u/2221746?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thinkverse", "html_url": "https://github.com/thinkverse", "followers_url": "https://api.github.com/users/thinkverse/followers", "following_url": "https://api.github.com/users/thinkverse/following{/other_user}", "gists_url": "https://api.github.com/users/thinkverse/gists{/gist_id}", "starred_url": "https://api.github.com/users/thinkverse/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thinkverse/subscriptions", "organizations_url": "https://api.github.com/users/thinkverse/orgs", "repos_url": "https://api.github.com/users/thinkverse/repos", "events_url": "https://api.github.com/users/thinkverse/events{/privacy}", "received_events_url": "https://api.github.com/users/thinkverse/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4109/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4109/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5009
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5009/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5009/comments
https://api.github.com/repos/ollama/ollama/issues/5009/events
https://github.com/ollama/ollama/pull/5009
2,349,769,196
PR_kwDOJ0Z1Ps5ySn5l
5,009
move OLLAMA_HOST to envconfig
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-12T22:17:39
2024-06-12T22:48:19
2024-06-12T22:48:16
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5009", "html_url": "https://github.com/ollama/ollama/pull/5009", "diff_url": "https://github.com/ollama/ollama/pull/5009.diff", "patch_url": "https://github.com/ollama/ollama/pull/5009.patch", "merged_at": "2024-06-12T22:48:16" }
null
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5009/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5009/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2927
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2927/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2927/comments
https://api.github.com/repos/ollama/ollama/issues/2927/events
https://github.com/ollama/ollama/pull/2927
2,168,268,598
PR_kwDOJ0Z1Ps5oqphV
2,927
Add Community Integration: OllamaGUI
{ "login": "enoch1118", "id": 54224095, "node_id": "MDQ6VXNlcjU0MjI0MDk1", "avatar_url": "https://avatars.githubusercontent.com/u/54224095?v=4", "gravatar_id": "", "url": "https://api.github.com/users/enoch1118", "html_url": "https://github.com/enoch1118", "followers_url": "https://api.github.com/users/enoch1118/followers", "following_url": "https://api.github.com/users/enoch1118/following{/other_user}", "gists_url": "https://api.github.com/users/enoch1118/gists{/gist_id}", "starred_url": "https://api.github.com/users/enoch1118/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/enoch1118/subscriptions", "organizations_url": "https://api.github.com/users/enoch1118/orgs", "repos_url": "https://api.github.com/users/enoch1118/repos", "events_url": "https://api.github.com/users/enoch1118/events{/privacy}", "received_events_url": "https://api.github.com/users/enoch1118/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-05T04:21:38
2024-03-25T18:58:29
2024-03-25T18:58:29
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2927", "html_url": "https://github.com/ollama/ollama/pull/2927", "diff_url": "https://github.com/ollama/ollama/pull/2927.diff", "patch_url": "https://github.com/ollama/ollama/pull/2927.patch", "merged_at": "2024-03-25T18:58:29" }
👋 I have added new integrations for macos! https://github.com/ollama/ollama/assets/54224095/eff7fd87-d8f6-4b97-bf06-f52e40e56dcc link: https://github.com/enoch1118/ollamaGUI
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2927/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2927/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6612
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6612/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6612/comments
https://api.github.com/repos/ollama/ollama/issues/6612/events
https://github.com/ollama/ollama/issues/6612
2,503,566,377
I_kwDOJ0Z1Ps6VOWQp
6,612
GPU Not Use
{ "login": "MEnsar55", "id": 124371091, "node_id": "U_kgDOB2nAkw", "avatar_url": "https://avatars.githubusercontent.com/u/124371091?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MEnsar55", "html_url": "https://github.com/MEnsar55", "followers_url": "https://api.github.com/users/MEnsar55/followers", "following_url": "https://api.github.com/users/MEnsar55/following{/other_user}", "gists_url": "https://api.github.com/users/MEnsar55/gists{/gist_id}", "starred_url": "https://api.github.com/users/MEnsar55/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MEnsar55/subscriptions", "organizations_url": "https://api.github.com/users/MEnsar55/orgs", "repos_url": "https://api.github.com/users/MEnsar55/repos", "events_url": "https://api.github.com/users/MEnsar55/events{/privacy}", "received_events_url": "https://api.github.com/users/MEnsar55/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-09-03T19:01:25
2024-09-25T22:34:25
2024-09-25T22:34:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hello, I am using Windows 11 and I have installed LLama3.1 7B and Gemma 2 27B in Ollama. When using these modems the GPU is between 0%-1%. What should I do? My laptop specs are: HP Victus 14700hx 4070 32 GB Ram
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6612/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6612/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2154
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2154/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2154/comments
https://api.github.com/repos/ollama/ollama/issues/2154/events
https://github.com/ollama/ollama/issues/2154
2,095,391,792
I_kwDOJ0Z1Ps585SQw
2,154
How to design our own prompt by import ollama?
{ "login": "haomes", "id": 82690723, "node_id": "MDQ6VXNlcjgyNjkwNzIz", "avatar_url": "https://avatars.githubusercontent.com/u/82690723?v=4", "gravatar_id": "", "url": "https://api.github.com/users/haomes", "html_url": "https://github.com/haomes", "followers_url": "https://api.github.com/users/haomes/followers", "following_url": "https://api.github.com/users/haomes/following{/other_user}", "gists_url": "https://api.github.com/users/haomes/gists{/gist_id}", "starred_url": "https://api.github.com/users/haomes/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/haomes/subscriptions", "organizations_url": "https://api.github.com/users/haomes/orgs", "repos_url": "https://api.github.com/users/haomes/repos", "events_url": "https://api.github.com/users/haomes/events{/privacy}", "received_events_url": "https://api.github.com/users/haomes/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-23T07:04:42
2024-03-12T22:54:26
2024-03-12T22:54:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
假如我想设计一个CR相关的prompt(比如:你是一个CR专家,请帮我根据提供的代码判断是否符标准...),并且通过 import ollama的python方法去调用大模型,我应该如何操作?
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2154/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2154/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5590
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5590/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5590/comments
https://api.github.com/repos/ollama/ollama/issues/5590/events
https://github.com/ollama/ollama/issues/5590
2,399,990,288
I_kwDOJ0Z1Ps6PDPIQ
5,590
Ollama running requests slow, while not utilizing entire VRAM
{ "login": "txhno", "id": 34884830, "node_id": "MDQ6VXNlcjM0ODg0ODMw", "avatar_url": "https://avatars.githubusercontent.com/u/34884830?v=4", "gravatar_id": "", "url": "https://api.github.com/users/txhno", "html_url": "https://github.com/txhno", "followers_url": "https://api.github.com/users/txhno/followers", "following_url": "https://api.github.com/users/txhno/following{/other_user}", "gists_url": "https://api.github.com/users/txhno/gists{/gist_id}", "starred_url": "https://api.github.com/users/txhno/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/txhno/subscriptions", "organizations_url": "https://api.github.com/users/txhno/orgs", "repos_url": "https://api.github.com/users/txhno/repos", "events_url": "https://api.github.com/users/txhno/events{/privacy}", "received_events_url": "https://api.github.com/users/txhno/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
4
2024-07-10T07:35:29
2024-07-26T18:17:53
2024-07-23T23:21:30
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I am continuously sending "generate" API requests to the Ollama server, querying the model wizardlm2:7b-q6_K. Each iteration takes approximately 3 seconds to respond. The model is using only 8GB of the 16GB VRAM available on my Tesla V100 GPU. Is there a way to make it utilize the entire VRAM to speed up request processing? If so, how can I achieve this?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5590/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5590/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4445
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4445/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4445/comments
https://api.github.com/repos/ollama/ollama/issues/4445/events
https://github.com/ollama/ollama/pull/4445
2,296,995,237
PR_kwDOJ0Z1Ps5ve7dU
4,445
doc: Add reference to package managers in Linux Installation section
{ "login": "tusharhero", "id": 54012021, "node_id": "MDQ6VXNlcjU0MDEyMDIx", "avatar_url": "https://avatars.githubusercontent.com/u/54012021?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tusharhero", "html_url": "https://github.com/tusharhero", "followers_url": "https://api.github.com/users/tusharhero/followers", "following_url": "https://api.github.com/users/tusharhero/following{/other_user}", "gists_url": "https://api.github.com/users/tusharhero/gists{/gist_id}", "starred_url": "https://api.github.com/users/tusharhero/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tusharhero/subscriptions", "organizations_url": "https://api.github.com/users/tusharhero/orgs", "repos_url": "https://api.github.com/users/tusharhero/repos", "events_url": "https://api.github.com/users/tusharhero/events{/privacy}", "received_events_url": "https://api.github.com/users/tusharhero/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-05-15T06:40:28
2024-11-25T00:00:21
2024-11-25T00:00:20
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4445", "html_url": "https://github.com/ollama/ollama/pull/4445", "diff_url": "https://github.com/ollama/ollama/pull/4445.diff", "patch_url": "https://github.com/ollama/ollama/pull/4445.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4445/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4445/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4277
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4277/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4277/comments
https://api.github.com/repos/ollama/ollama/issues/4277/events
https://github.com/ollama/ollama/issues/4277
2,287,108,723
I_kwDOJ0Z1Ps6IUoJz
4,277
Unexpected Increase in Inference Time as Context Window Grows on Llama3:7b
{ "login": "gusanmaz", "id": 2552975, "node_id": "MDQ6VXNlcjI1NTI5NzU=", "avatar_url": "https://avatars.githubusercontent.com/u/2552975?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gusanmaz", "html_url": "https://github.com/gusanmaz", "followers_url": "https://api.github.com/users/gusanmaz/followers", "following_url": "https://api.github.com/users/gusanmaz/following{/other_user}", "gists_url": "https://api.github.com/users/gusanmaz/gists{/gist_id}", "starred_url": "https://api.github.com/users/gusanmaz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gusanmaz/subscriptions", "organizations_url": "https://api.github.com/users/gusanmaz/orgs", "repos_url": "https://api.github.com/users/gusanmaz/repos", "events_url": "https://api.github.com/users/gusanmaz/events{/privacy}", "received_events_url": "https://api.github.com/users/gusanmaz/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5808482718, "node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng", "url": "https://api.github.com/repos/ollama/ollama/labels/performance", "name": "performance", "color": "A5B5C6", "default": false, "description": "" } ]
closed
false
null
[]
null
3
2024-05-09T07:49:19
2024-10-23T21:08:16
2024-10-23T21:08:16
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am doing some benchmarks on RAG using llama3:7b model on Ollama. I ask a question first directly to the model, then ask the question and provide context from relevant documents, asking the model to answer the question based on the given context; essentially, asking the question in the RAG way without exceeding the model's context window. As expected, the first query is a sentence long, and the second query is many sentences long. I asked 14 questions (14 queries for direct questions and 14 queries for RAG questions in total for each machine) and the benchmark results can be seen below: | Machine Type | CPU | RAM (GB) | Graphics Card | OS | Direct Question - Short Context (ms) | RAG Question - Long Context (ms) | |----------------|----------------------------------------|----------|--------------------------------------------------|------------------|------------|----------| | Mac Mini | Apple Silicon M2 Pro | 16 | | macOS 14.2.1 | 61152 | 105998 | | Laptop | AMD Ryzen 9 5900HX (16) @ 4.680GHz | 32 | NVIDIA GeForce RTX 3050 Mobile, AMD ATI Cezanne | Pop!_OS 22.04 LTS| 413264 | 1052304 | | Desktop | 11th Gen Intel i5-11400F (12) @ 4.400GHz| 64 | NVIDIA GeForce RTX 3060 Lite Hash Rate | Pop!_OS 22.04 LTS| 114599 | 152341 | I use Ollama version 0.1.34. As far as I know, inference time doesn't change significantly as query context grows for LLMs. I am particularly surprised to see more than a 2.5x increase in inference time on my laptop machine. I haven't performed this benchmark on a different model than Llama3. I wonder if something is wrong with Ollama or if these benchmark results I am getting are normal. Thanks! ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.1.34
{ "login": "jessegross", "id": 6468499, "node_id": "MDQ6VXNlcjY0Njg0OTk=", "avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jessegross", "html_url": "https://github.com/jessegross", "followers_url": "https://api.github.com/users/jessegross/followers", "following_url": "https://api.github.com/users/jessegross/following{/other_user}", "gists_url": "https://api.github.com/users/jessegross/gists{/gist_id}", "starred_url": "https://api.github.com/users/jessegross/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jessegross/subscriptions", "organizations_url": "https://api.github.com/users/jessegross/orgs", "repos_url": "https://api.github.com/users/jessegross/repos", "events_url": "https://api.github.com/users/jessegross/events{/privacy}", "received_events_url": "https://api.github.com/users/jessegross/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4277/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4277/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8281
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8281/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8281/comments
https://api.github.com/repos/ollama/ollama/issues/8281/events
https://github.com/ollama/ollama/issues/8281
2,765,168,502
I_kwDOJ0Z1Ps6k0R92
8,281
Runing ollama on Intel Ultra NPU or GPU
{ "login": "jackphj", "id": 17948210, "node_id": "MDQ6VXNlcjE3OTQ4MjEw", "avatar_url": "https://avatars.githubusercontent.com/u/17948210?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jackphj", "html_url": "https://github.com/jackphj", "followers_url": "https://api.github.com/users/jackphj/followers", "following_url": "https://api.github.com/users/jackphj/following{/other_user}", "gists_url": "https://api.github.com/users/jackphj/gists{/gist_id}", "starred_url": "https://api.github.com/users/jackphj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jackphj/subscriptions", "organizations_url": "https://api.github.com/users/jackphj/orgs", "repos_url": "https://api.github.com/users/jackphj/repos", "events_url": "https://api.github.com/users/jackphj/events{/privacy}", "received_events_url": "https://api.github.com/users/jackphj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 6677491450, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgJu-g", "url": "https://api.github.com/repos/ollama/ollama/labels/intel", "name": "intel", "color": "226E5B", "default": false, "description": "issues relating to Intel GPUs" }, { "id": 6677745918, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g", "url": "https://api.github.com/repos/ollama/ollama/labels/gpu", "name": "gpu", "color": "76C49E", "default": false, "description": "" } ]
open
false
null
[]
null
7
2025-01-01T17:04:56
2025-01-24T12:49:19
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
After I installed ollama through ollamaSetup, I found that it cannot use my gpu or npu. How to solve this problem? CPU: intel ultra7 258v System: windows 11 24h2
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8281/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8281/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/57
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/57/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/57/comments
https://api.github.com/repos/ollama/ollama/issues/57/events
https://github.com/ollama/ollama/pull/57
1,794,211,902
PR_kwDOJ0Z1Ps5U9BQy
57
take all args as one prompt
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-07-07T20:16:27
2023-07-11T14:19:36
2023-07-10T10:05:09
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/57", "html_url": "https://github.com/ollama/ollama/pull/57", "diff_url": "https://github.com/ollama/ollama/pull/57.diff", "patch_url": "https://github.com/ollama/ollama/pull/57.patch", "merged_at": "2023-07-10T10:05:09" }
- parse all run arguments into one prompt - do not echo prompt back on one-shot - example of summarizing a document ``` $ ollama run nous-hermes "$(cat input.txt)", please summarize this story The song "Summertime" by Will Smith is about the rapper's life before fame and how it changed when he became successful. The lyrics describe his experiences growing up in Philadelphia and then suddenly becoming rich and famous. The song is a coming-of-age story that describes the rapper's journey from poverty to stardom, and the impact of this change on his life. ```
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/57/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/57/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4597
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4597/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4597/comments
https://api.github.com/repos/ollama/ollama/issues/4597/events
https://github.com/ollama/ollama/pull/4597
2,313,575,748
PR_kwDOJ0Z1Ps5wXeYa
4,597
bump
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-23T18:23:06
2024-05-23T21:16:26
2024-05-23T21:16:26
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4597", "html_url": "https://github.com/ollama/ollama/pull/4597", "diff_url": "https://github.com/ollama/ollama/pull/4597.diff", "patch_url": "https://github.com/ollama/ollama/pull/4597.patch", "merged_at": "2024-05-23T21:16:26" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4597/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4597/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1208
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1208/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1208/comments
https://api.github.com/repos/ollama/ollama/issues/1208/events
https://github.com/ollama/ollama/pull/1208
2,002,491,049
PR_kwDOJ0Z1Ps5f7Peh
1,208
Add Laravel package to README.md
{ "login": "cloudstudio", "id": 3589377, "node_id": "MDQ6VXNlcjM1ODkzNzc=", "avatar_url": "https://avatars.githubusercontent.com/u/3589377?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cloudstudio", "html_url": "https://github.com/cloudstudio", "followers_url": "https://api.github.com/users/cloudstudio/followers", "following_url": "https://api.github.com/users/cloudstudio/following{/other_user}", "gists_url": "https://api.github.com/users/cloudstudio/gists{/gist_id}", "starred_url": "https://api.github.com/users/cloudstudio/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cloudstudio/subscriptions", "organizations_url": "https://api.github.com/users/cloudstudio/orgs", "repos_url": "https://api.github.com/users/cloudstudio/repos", "events_url": "https://api.github.com/users/cloudstudio/events{/privacy}", "received_events_url": "https://api.github.com/users/cloudstudio/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-11-20T15:46:16
2023-11-20T15:48:35
2023-11-20T15:48:35
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1208", "html_url": "https://github.com/ollama/ollama/pull/1208", "diff_url": "https://github.com/ollama/ollama/pull/1208.diff", "patch_url": "https://github.com/ollama/ollama/pull/1208.patch", "merged_at": "2023-11-20T15:48:35" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1208/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1208/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8026
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8026/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8026/comments
https://api.github.com/repos/ollama/ollama/issues/8026/events
https://github.com/ollama/ollama/issues/8026
2,730,023,228
I_kwDOJ0Z1Ps6iuNk8
8,026
OpenAI Chat Completion Client For Multimodal
{ "login": "iejzh", "id": 32932199, "node_id": "MDQ6VXNlcjMyOTMyMTk5", "avatar_url": "https://avatars.githubusercontent.com/u/32932199?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iejzh", "html_url": "https://github.com/iejzh", "followers_url": "https://api.github.com/users/iejzh/followers", "following_url": "https://api.github.com/users/iejzh/following{/other_user}", "gists_url": "https://api.github.com/users/iejzh/gists{/gist_id}", "starred_url": "https://api.github.com/users/iejzh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iejzh/subscriptions", "organizations_url": "https://api.github.com/users/iejzh/orgs", "repos_url": "https://api.github.com/users/iejzh/repos", "events_url": "https://api.github.com/users/iejzh/events{/privacy}", "received_events_url": "https://api.github.com/users/iejzh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q", "url": "https://api.github.com/repos/ollama/ollama/labels/api", "name": "api", "color": "bfdadc", "default": false, "description": "" } ]
closed
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-12-10T12:42:53
2024-12-26T05:02:37
2024-12-26T05:02:37
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Inconsistency with openai standard api parameters ```python response = client.chat.completions.create( model="gpt-4o-mini", messages=[ { "role": "user", "content": [ { "type": "text", "text": "What is in this image?", }, { "type": "image_url", "image_url": { "url": f"data:image/jpeg;base64,{base64_image}" }, }, ], } ], ) ```
{ "login": "iejzh", "id": 32932199, "node_id": "MDQ6VXNlcjMyOTMyMTk5", "avatar_url": "https://avatars.githubusercontent.com/u/32932199?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iejzh", "html_url": "https://github.com/iejzh", "followers_url": "https://api.github.com/users/iejzh/followers", "following_url": "https://api.github.com/users/iejzh/following{/other_user}", "gists_url": "https://api.github.com/users/iejzh/gists{/gist_id}", "starred_url": "https://api.github.com/users/iejzh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iejzh/subscriptions", "organizations_url": "https://api.github.com/users/iejzh/orgs", "repos_url": "https://api.github.com/users/iejzh/repos", "events_url": "https://api.github.com/users/iejzh/events{/privacy}", "received_events_url": "https://api.github.com/users/iejzh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8026/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8026/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8154
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8154/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8154/comments
https://api.github.com/repos/ollama/ollama/issues/8154/events
https://github.com/ollama/ollama/issues/8154
2,747,425,045
I_kwDOJ0Z1Ps6jwmEV
8,154
I can not connect to 11434 port
{ "login": "1760842797", "id": 144416737, "node_id": "U_kgDOCJuf4Q", "avatar_url": "https://avatars.githubusercontent.com/u/144416737?v=4", "gravatar_id": "", "url": "https://api.github.com/users/1760842797", "html_url": "https://github.com/1760842797", "followers_url": "https://api.github.com/users/1760842797/followers", "following_url": "https://api.github.com/users/1760842797/following{/other_user}", "gists_url": "https://api.github.com/users/1760842797/gists{/gist_id}", "starred_url": "https://api.github.com/users/1760842797/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/1760842797/subscriptions", "organizations_url": "https://api.github.com/users/1760842797/orgs", "repos_url": "https://api.github.com/users/1760842797/repos", "events_url": "https://api.github.com/users/1760842797/events{/privacy}", "received_events_url": "https://api.github.com/users/1760842797/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
open
false
null
[]
null
1
2024-12-18T10:30:19
2024-12-19T19:56:10
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? sir, i add Environment="OLLAMA_HOST=0.0.0.0" to the service file,Then i restart ollama. now, i find the service listener port is only have tcp6,without tcp4 listen,so i can not access to the service from other machine. how can i fix this problem? thank you [root@localhost ollama]# netstat -tuln | grep 11434 tcp6 0 0 :::11434 :::* LISTEN [root@localhost ollama]# ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8154/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8154/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3960
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3960/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3960/comments
https://api.github.com/repos/ollama/ollama/issues/3960/events
https://github.com/ollama/ollama/pull/3960
2,266,528,424
PR_kwDOJ0Z1Ps5t4gfb
3,960
server: add dynamic configuration for download variables
{ "login": "jzila", "id": 59594, "node_id": "MDQ6VXNlcjU5NTk0", "avatar_url": "https://avatars.githubusercontent.com/u/59594?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jzila", "html_url": "https://github.com/jzila", "followers_url": "https://api.github.com/users/jzila/followers", "following_url": "https://api.github.com/users/jzila/following{/other_user}", "gists_url": "https://api.github.com/users/jzila/gists{/gist_id}", "starred_url": "https://api.github.com/users/jzila/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jzila/subscriptions", "organizations_url": "https://api.github.com/users/jzila/orgs", "repos_url": "https://api.github.com/users/jzila/repos", "events_url": "https://api.github.com/users/jzila/events{/privacy}", "received_events_url": "https://api.github.com/users/jzila/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2024-04-26T21:27:46
2024-11-21T14:47:44
2024-11-21T09:56:35
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3960", "html_url": "https://github.com/ollama/ollama/pull/3960", "diff_url": "https://github.com/ollama/ollama/pull/3960.diff", "patch_url": "https://github.com/ollama/ollama/pull/3960.patch", "merged_at": null }
## Issue I've found that downloads can be unreliable for large models, either due to errors during the download or during processing the digest. ## Workaround Add some environment variables to the server to allow runtime configuration of the number of parallel downloads and how big the chunks can be. ## Follow-up Ideally the server would be configured to respect the `?partNumber` query parameters that the returned `x-amz-mp-parts-count` header implies should be supported. I'd also like it to support `x-amz-checksum-mode=ENABLED` (currently returns a `501 NOT IMPLEMENTED`), so that each part number returns an expected digest in the response headers for `GET` and `HEAD`. This would enable us to split the digest by parts, so that if a part download fails, we don't need to retrieve the full model each time. It seems that the current implementation is done via [github.com/distribution/distribution](github.com/distribution/distribution), which delegates its `Range` handling to `http.ServeContent`, and as such does not support the `partNumber` functionality implied by the CloudFlare response headers. So any such support would have to fork github.com/distribution/distribution. CloudFlare also does not appear to support these headers.
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3960/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3960/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7990
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7990/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7990/comments
https://api.github.com/repos/ollama/ollama/issues/7990/events
https://github.com/ollama/ollama/issues/7990
2,724,838,177
I_kwDOJ0Z1Ps6iabsh
7,990
Incorrect version for v0.5.1 - ollama version is 0.4.4
{ "login": "vt-alt", "id": 36664211, "node_id": "MDQ6VXNlcjM2NjY0MjEx", "avatar_url": "https://avatars.githubusercontent.com/u/36664211?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vt-alt", "html_url": "https://github.com/vt-alt", "followers_url": "https://api.github.com/users/vt-alt/followers", "following_url": "https://api.github.com/users/vt-alt/following{/other_user}", "gists_url": "https://api.github.com/users/vt-alt/gists{/gist_id}", "starred_url": "https://api.github.com/users/vt-alt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vt-alt/subscriptions", "organizations_url": "https://api.github.com/users/vt-alt/orgs", "repos_url": "https://api.github.com/users/vt-alt/repos", "events_url": "https://api.github.com/users/vt-alt/events{/privacy}", "received_events_url": "https://api.github.com/users/vt-alt/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-12-07T22:27:56
2024-12-07T22:28:57
2024-12-07T22:28:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? `ollama` binary from ollama-linux-amd64.tgz from https://github.com/ollama/ollama/releases/tag/v0.5.1 shows incorrect version 0.4.4. ``` $ bin/ollama --version ollama version is 0.4.4 Warning: client version is 0.5.1 ``` ### OS Linux ### GPU Intel ### CPU Intel ### Ollama version v0.5.1
{ "login": "vt-alt", "id": 36664211, "node_id": "MDQ6VXNlcjM2NjY0MjEx", "avatar_url": "https://avatars.githubusercontent.com/u/36664211?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vt-alt", "html_url": "https://github.com/vt-alt", "followers_url": "https://api.github.com/users/vt-alt/followers", "following_url": "https://api.github.com/users/vt-alt/following{/other_user}", "gists_url": "https://api.github.com/users/vt-alt/gists{/gist_id}", "starred_url": "https://api.github.com/users/vt-alt/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vt-alt/subscriptions", "organizations_url": "https://api.github.com/users/vt-alt/orgs", "repos_url": "https://api.github.com/users/vt-alt/repos", "events_url": "https://api.github.com/users/vt-alt/events{/privacy}", "received_events_url": "https://api.github.com/users/vt-alt/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7990/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7990/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4241
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4241/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4241/comments
https://api.github.com/repos/ollama/ollama/issues/4241/events
https://github.com/ollama/ollama/pull/4241
2,284,406,156
PR_kwDOJ0Z1Ps5u0XnE
4,241
Detect noexec and report a better error
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-07T23:49:02
2024-05-08T22:34:25
2024-05-08T22:34:22
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4241", "html_url": "https://github.com/ollama/ollama/pull/4241", "diff_url": "https://github.com/ollama/ollama/pull/4241.diff", "patch_url": "https://github.com/ollama/ollama/pull/4241.patch", "merged_at": "2024-05-08T22:34:22" }
This will bubble up a much more informative error message if noexec is preventing us from running the subprocess Fixes #4105 Repro scenario: ``` sudo mkdir /tmp2 sudo mount -t tmpfs -o size=1g,noexec tmpfs /tmp2 OLLAMA_TMPDIR=/tmp2/test1 ollama serve ... time=2024-05-07T23:44:54.304Z level=INFO source=sched.go:308 msg="NewLlamaServer failed" model=/home/daniel/.ollama/models/blobs/sha256-66002b78c70a22ab25e16cc9a1736c6cc6335398c7312e3eb33db202350afe66 error="unable to start server fork/exec /tmp2/test1/runners/cpu_avx/ollama_llama_server: permission denied. /tmp2/test1/runners/cpu_avx may have noexec set. Set OLLAMA_TMPDIR for server to a writable executable directory" ``` On the client that triggered the failure: ``` % ollama run orca-mini hello Error: unable to start server fork/exec /tmp2/test1/runners/cpu_avx/ollama_llama_server: permission denied. /tmp2/test1/runners/cpu_avx may have noexec set. Set OLLAMA_TMPDIR for server to a writable executable directory ```
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4241/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4241/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1195
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1195/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1195/comments
https://api.github.com/repos/ollama/ollama/issues/1195/events
https://github.com/ollama/ollama/pull/1195
2,000,669,275
PR_kwDOJ0Z1Ps5f1L1O
1,195
progress: fix bar rate
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
3
2023-11-19T02:37:58
2023-11-28T19:55:24
2023-11-28T19:55:23
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1195", "html_url": "https://github.com/ollama/ollama/pull/1195", "diff_url": "https://github.com/ollama/ollama/pull/1195.diff", "patch_url": "https://github.com/ollama/ollama/pull/1195.patch", "merged_at": "2023-11-28T19:55:23" }
implement rate as a rolling average over the last n updates. the current issue is rate is calculated as an average rate over the lifetime of the progress bar. this reflects the actual progress well if the progress is smooth and flat but that's rarely the case. the rolling average is a better way to represent changing rates. if more progress is made in the window than before, the rate will go up. if less progress is made, the rate will go down
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1195/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1195/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5863
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5863/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5863/comments
https://api.github.com/repos/ollama/ollama/issues/5863/events
https://github.com/ollama/ollama/issues/5863
2,423,893,740
I_kwDOJ0Z1Ps6Qea7s
5,863
function calling documentation
{ "login": "UmutAlihan", "id": 16688836, "node_id": "MDQ6VXNlcjE2Njg4ODM2", "avatar_url": "https://avatars.githubusercontent.com/u/16688836?v=4", "gravatar_id": "", "url": "https://api.github.com/users/UmutAlihan", "html_url": "https://github.com/UmutAlihan", "followers_url": "https://api.github.com/users/UmutAlihan/followers", "following_url": "https://api.github.com/users/UmutAlihan/following{/other_user}", "gists_url": "https://api.github.com/users/UmutAlihan/gists{/gist_id}", "starred_url": "https://api.github.com/users/UmutAlihan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/UmutAlihan/subscriptions", "organizations_url": "https://api.github.com/users/UmutAlihan/orgs", "repos_url": "https://api.github.com/users/UmutAlihan/repos", "events_url": "https://api.github.com/users/UmutAlihan/events{/privacy}", "received_events_url": "https://api.github.com/users/UmutAlihan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396191, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aXw", "url": "https://api.github.com/repos/ollama/ollama/labels/documentation", "name": "documentation", "color": "0075ca", "default": true, "description": "Improvements or additions to documentation" } ]
closed
false
null
[]
null
7
2024-07-22T22:50:09
2024-07-26T19:36:38
2024-07-26T00:45:56
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
While searching the web I couldn't yet find a resource except function callilng through mistral_inference package. I am curious whether anyone wrote a raw prompt in which function calling is working?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5863/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5863/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5867
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5867/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5867/comments
https://api.github.com/repos/ollama/ollama/issues/5867/events
https://github.com/ollama/ollama/issues/5867
2,424,285,265
I_kwDOJ0Z1Ps6Qf6hR
5,867
Nvidia Minitron Please!
{ "login": "txhno", "id": 34884830, "node_id": "MDQ6VXNlcjM0ODg0ODMw", "avatar_url": "https://avatars.githubusercontent.com/u/34884830?v=4", "gravatar_id": "", "url": "https://api.github.com/users/txhno", "html_url": "https://github.com/txhno", "followers_url": "https://api.github.com/users/txhno/followers", "following_url": "https://api.github.com/users/txhno/following{/other_user}", "gists_url": "https://api.github.com/users/txhno/gists{/gist_id}", "starred_url": "https://api.github.com/users/txhno/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/txhno/subscriptions", "organizations_url": "https://api.github.com/users/txhno/orgs", "repos_url": "https://api.github.com/users/txhno/repos", "events_url": "https://api.github.com/users/txhno/events{/privacy}", "received_events_url": "https://api.github.com/users/txhno/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2024-07-23T05:25:05
2024-09-10T17:12:18
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://huggingface.co/collections/nvidia/minitron-669ac727dc9c86e6ab7f0f3e
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5867/reactions", "total_count": 14, "+1": 12, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5867/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4384
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4384/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4384/comments
https://api.github.com/repos/ollama/ollama/issues/4384/events
https://github.com/ollama/ollama/issues/4384
2,291,538,755
I_kwDOJ0Z1Ps6IlhtD
4,384
Ollama "ollama show --modelfile" shows license instead of modelfile
{ "login": "omegatao", "id": 11746995, "node_id": "MDQ6VXNlcjExNzQ2OTk1", "avatar_url": "https://avatars.githubusercontent.com/u/11746995?v=4", "gravatar_id": "", "url": "https://api.github.com/users/omegatao", "html_url": "https://github.com/omegatao", "followers_url": "https://api.github.com/users/omegatao/followers", "following_url": "https://api.github.com/users/omegatao/following{/other_user}", "gists_url": "https://api.github.com/users/omegatao/gists{/gist_id}", "starred_url": "https://api.github.com/users/omegatao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/omegatao/subscriptions", "organizations_url": "https://api.github.com/users/omegatao/orgs", "repos_url": "https://api.github.com/users/omegatao/repos", "events_url": "https://api.github.com/users/omegatao/events{/privacy}", "received_events_url": "https://api.github.com/users/omegatao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-05-12T23:02:26
2024-05-14T09:35:14
2024-05-14T06:34:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When the command 'ollama show --modelfile llama3:70b' is entered, it should have displayed information about the model's modelfile, but instead, it displayed license information. The "--modelfile: parameter for "ollama show" command seems not working. ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.1.36
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4384/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4384/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/403
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/403/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/403/comments
https://api.github.com/repos/ollama/ollama/issues/403/events
https://github.com/ollama/ollama/issues/403
1,864,427,115
I_kwDOJ0Z1Ps5vIOZr
403
Ollama Windows version
{ "login": "deadcoder0904", "id": 16436270, "node_id": "MDQ6VXNlcjE2NDM2Mjcw", "avatar_url": "https://avatars.githubusercontent.com/u/16436270?v=4", "gravatar_id": "", "url": "https://api.github.com/users/deadcoder0904", "html_url": "https://github.com/deadcoder0904", "followers_url": "https://api.github.com/users/deadcoder0904/followers", "following_url": "https://api.github.com/users/deadcoder0904/following{/other_user}", "gists_url": "https://api.github.com/users/deadcoder0904/gists{/gist_id}", "starred_url": "https://api.github.com/users/deadcoder0904/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/deadcoder0904/subscriptions", "organizations_url": "https://api.github.com/users/deadcoder0904/orgs", "repos_url": "https://api.github.com/users/deadcoder0904/repos", "events_url": "https://api.github.com/users/deadcoder0904/events{/privacy}", "received_events_url": "https://api.github.com/users/deadcoder0904/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
36
2023-08-24T05:23:02
2024-05-23T11:31:34
2024-02-16T16:21:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I saw it is coming but didn't mention when? Would be great if you pinned this issue as more people use Windows & ollama has such a great dx. The project looks absolutely brilliant. Would love to use text (gpt-4) & code (copilot) locally.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/403/reactions", "total_count": 33, "+1": 24, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 4, "rocket": 0, "eyes": 5 }
https://api.github.com/repos/ollama/ollama/issues/403/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3485
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3485/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3485/comments
https://api.github.com/repos/ollama/ollama/issues/3485/events
https://github.com/ollama/ollama/issues/3485
2,225,203,137
I_kwDOJ0Z1Ps6EoefB
3,485
Whether Qwen from ollama library is Qwen1.5?
{ "login": "OPDEV001", "id": 120762872, "node_id": "U_kgDOBzKx-A", "avatar_url": "https://avatars.githubusercontent.com/u/120762872?v=4", "gravatar_id": "", "url": "https://api.github.com/users/OPDEV001", "html_url": "https://github.com/OPDEV001", "followers_url": "https://api.github.com/users/OPDEV001/followers", "following_url": "https://api.github.com/users/OPDEV001/following{/other_user}", "gists_url": "https://api.github.com/users/OPDEV001/gists{/gist_id}", "starred_url": "https://api.github.com/users/OPDEV001/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/OPDEV001/subscriptions", "organizations_url": "https://api.github.com/users/OPDEV001/orgs", "repos_url": "https://api.github.com/users/OPDEV001/repos", "events_url": "https://api.github.com/users/OPDEV001/events{/privacy}", "received_events_url": "https://api.github.com/users/OPDEV001/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-04-04T11:19:33
2024-04-23T23:20:56
2024-04-17T22:02:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What model would you like? When I type ollama run qwen:7b and the command completed successfully, but may I know whether the Qwen is Qwen or Qwen1.5, because I saw Qwen1.5 in Github. If not, please add Qwen1.5 to ollama library. Thanks,
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3485/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3485/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6917
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6917/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6917/comments
https://api.github.com/repos/ollama/ollama/issues/6917/events
https://github.com/ollama/ollama/issues/6917
2,542,590,463
I_kwDOJ0Z1Ps6XjNn_
6,917
downloadChunk does not pass the Authorization header to the registry
{ "login": "alexmavr", "id": 680441, "node_id": "MDQ6VXNlcjY4MDQ0MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/680441?v=4", "gravatar_id": "", "url": "https://api.github.com/users/alexmavr", "html_url": "https://github.com/alexmavr", "followers_url": "https://api.github.com/users/alexmavr/followers", "following_url": "https://api.github.com/users/alexmavr/following{/other_user}", "gists_url": "https://api.github.com/users/alexmavr/gists{/gist_id}", "starred_url": "https://api.github.com/users/alexmavr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/alexmavr/subscriptions", "organizations_url": "https://api.github.com/users/alexmavr/orgs", "repos_url": "https://api.github.com/users/alexmavr/repos", "events_url": "https://api.github.com/users/alexmavr/events{/privacy}", "received_events_url": "https://api.github.com/users/alexmavr/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-09-23T13:09:23
2024-10-05T21:54:04
null
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? In #5994 , `regOpts` were removed from the `blobDownload.downloadChunk` method as unnecessary. While it's true that the ollama library only offers public blobs, the lack of regOpts means that all `GET /v2/<image>/blobs/...` requests with a `Range` header cannot be accompanied by an `Authorization` header. This removal now breaks the mirrored behavior with [blobUpload.uploadPart](https://github.com/ollama/ollama/blob/main/server/upload.go#L152) where the RegOpts and corresponding `Authorization header` are still passed to each invocation. I tried adding the `Authorization: Bearer` token to the request, but it looks like pulls from `registry.ollama.ai` are breaking with this change, most likely because the header is being passed to the cloudflarestorage.com URLs. ``` ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.3.11
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6917/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6917/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8668
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8668/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8668/comments
https://api.github.com/repos/ollama/ollama/issues/8668/events
https://github.com/ollama/ollama/pull/8668
2,818,695,620
PR_kwDOJ0Z1Ps6JYzfR
8,668
Hide empty terminal window
{ "login": "ashokgelal", "id": 401055, "node_id": "MDQ6VXNlcjQwMTA1NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/401055?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ashokgelal", "html_url": "https://github.com/ashokgelal", "followers_url": "https://api.github.com/users/ashokgelal/followers", "following_url": "https://api.github.com/users/ashokgelal/following{/other_user}", "gists_url": "https://api.github.com/users/ashokgelal/gists{/gist_id}", "starred_url": "https://api.github.com/users/ashokgelal/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ashokgelal/subscriptions", "organizations_url": "https://api.github.com/users/ashokgelal/orgs", "repos_url": "https://api.github.com/users/ashokgelal/repos", "events_url": "https://api.github.com/users/ashokgelal/events{/privacy}", "received_events_url": "https://api.github.com/users/ashokgelal/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2025-01-29T16:31:07
2025-01-29T16:31:07
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8668", "html_url": "https://github.com/ollama/ollama/pull/8668", "diff_url": "https://github.com/ollama/ollama/pull/8668.diff", "patch_url": "https://github.com/ollama/ollama/pull/8668.patch", "merged_at": null }
This hides the LlamaServer blank window when chatting outside of the terminal (say like with an app like Msty). This has no other side effects when invoking it the regular way. I had sent a PR for this a while ago and it was closed thinking it had been resolved but this issue still exists. (see: https://github.com/ollama/ollama/pull/4287)
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8668/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8668/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3732
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3732/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3732/comments
https://api.github.com/repos/ollama/ollama/issues/3732/events
https://github.com/ollama/ollama/issues/3732
2,250,612,421
I_kwDOJ0Z1Ps6GJZ7F
3,732
num_gpu is not working in modelfile based on another model.
{ "login": "chigkim", "id": 22120994, "node_id": "MDQ6VXNlcjIyMTIwOTk0", "avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chigkim", "html_url": "https://github.com/chigkim", "followers_url": "https://api.github.com/users/chigkim/followers", "following_url": "https://api.github.com/users/chigkim/following{/other_user}", "gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}", "starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chigkim/subscriptions", "organizations_url": "https://api.github.com/users/chigkim/orgs", "repos_url": "https://api.github.com/users/chigkim/repos", "events_url": "https://api.github.com/users/chigkim/events{/privacy}", "received_events_url": "https://api.github.com/users/chigkim/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-04-18T12:48:15
2024-07-11T03:46:19
2024-07-11T03:46:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? First I downloaded wizardlm2:8x22b. `ollama pull wizardlm2:8x22b` I'm trying to offload only 30 layers to gpu using this modelfile with PARAMETER num_gpu 30: ``` FROM wizardlm2:8x22b TEMPLATE """{{ if .System }}{{ .System }} {{ end }}{{ if .Prompt }}USER: {{ .Prompt }} {{ end }}ASSISTANT: {{ .Response }}""" SYSTEM """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.""" PARAMETER stop "USER:" PARAMETER stop "ASSISTANT:" PARAMETER num_gpu 30 ``` However, server.log indicates 0 layers are offloaded. If I create a model from .gguf file that I downloaded from HF, it works. ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.3.2
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3732/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3732/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4304
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4304/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4304/comments
https://api.github.com/repos/ollama/ollama/issues/4304/events
https://github.com/ollama/ollama/pull/4304
2,288,578,867
PR_kwDOJ0Z1Ps5vCazn
4,304
Fix race in shutdown logic
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-09T22:48:49
2024-05-09T22:58:58
2024-05-09T22:58:45
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4304", "html_url": "https://github.com/ollama/ollama/pull/4304", "diff_url": "https://github.com/ollama/ollama/pull/4304.diff", "patch_url": "https://github.com/ollama/ollama/pull/4304.patch", "merged_at": "2024-05-09T22:58:45" }
Ensure the runners are terminated Fixes #4267
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4304/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4304/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4594
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4594/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4594/comments
https://api.github.com/repos/ollama/ollama/issues/4594/events
https://github.com/ollama/ollama/pull/4594
2,313,383,454
PR_kwDOJ0Z1Ps5wWz9k
4,594
Add isolated gpu test to troubleshooting
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-23T16:34:41
2024-05-30T20:10:57
2024-05-30T20:10:55
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4594", "html_url": "https://github.com/ollama/ollama/pull/4594", "diff_url": "https://github.com/ollama/ollama/pull/4594.diff", "patch_url": "https://github.com/ollama/ollama/pull/4594.patch", "merged_at": "2024-05-30T20:10:55" }
This should help users isolate problems with their container runtime.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4594/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4594/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8393
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8393/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8393/comments
https://api.github.com/repos/ollama/ollama/issues/8393/events
https://github.com/ollama/ollama/issues/8393
2,782,326,999
I_kwDOJ0Z1Ps6l1vDX
8,393
Unable to enable GPU for models
{ "login": "hyongaa", "id": 145748303, "node_id": "U_kgDOCK_xTw", "avatar_url": "https://avatars.githubusercontent.com/u/145748303?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hyongaa", "html_url": "https://github.com/hyongaa", "followers_url": "https://api.github.com/users/hyongaa/followers", "following_url": "https://api.github.com/users/hyongaa/following{/other_user}", "gists_url": "https://api.github.com/users/hyongaa/gists{/gist_id}", "starred_url": "https://api.github.com/users/hyongaa/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hyongaa/subscriptions", "organizations_url": "https://api.github.com/users/hyongaa/orgs", "repos_url": "https://api.github.com/users/hyongaa/repos", "events_url": "https://api.github.com/users/hyongaa/events{/privacy}", "received_events_url": "https://api.github.com/users/hyongaa/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
16
2025-01-12T07:18:05
2025-01-16T00:37:54
2025-01-16T00:37:54
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I don't have much knowledge in this kind of stuff, so I followed the instructions and installed ollama. But when I run the model, only CPU is working and my nvidia GPU doesn't work at all. I am wondering how can I enable GPU for ollama. Thanks a lot. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.4
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8393/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8393/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8652
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8652/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8652/comments
https://api.github.com/repos/ollama/ollama/issues/8652/events
https://github.com/ollama/ollama/issues/8652
2,817,696,992
I_kwDOJ0Z1Ps6n8qTg
8,652
Could not install deepseek-r1:7b version
{ "login": "chinmayc0707", "id": 139462368, "node_id": "U_kgDOCFAG4A", "avatar_url": "https://avatars.githubusercontent.com/u/139462368?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chinmayc0707", "html_url": "https://github.com/chinmayc0707", "followers_url": "https://api.github.com/users/chinmayc0707/followers", "following_url": "https://api.github.com/users/chinmayc0707/following{/other_user}", "gists_url": "https://api.github.com/users/chinmayc0707/gists{/gist_id}", "starred_url": "https://api.github.com/users/chinmayc0707/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chinmayc0707/subscriptions", "organizations_url": "https://api.github.com/users/chinmayc0707/orgs", "repos_url": "https://api.github.com/users/chinmayc0707/repos", "events_url": "https://api.github.com/users/chinmayc0707/events{/privacy}", "received_events_url": "https://api.github.com/users/chinmayc0707/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677370291, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw", "url": "https://api.github.com/repos/ollama/ollama/labels/networking", "name": "networking", "color": "0B5368", "default": false, "description": "Issues relating to ollama pull and push" } ]
open
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2025-01-29T09:42:18
2025-01-30T08:54:15
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Error: max retries exceeded: Get "https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ollama/docker/registry/v2/blobs/sha256/aa/aabd4debf0c8f08881923f2c25fc0fdeed24435271c2b3e92c4af36704040dbc/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=66040c77ac1b787c3af820529859349a%2F20250129%2Fauto%2Fs3%2Faws4_request&X-Amz-Date=20250129T093708Z&X-Amz-Expires=86400&X-Amz-SignedHeaders=host&X-Amz-Signature=044eb3053c34ac7e27101080923c27a8d4f0adc5334de93463391b2cf3d208cd": dial tcp: lookup dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com: no such host ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8652/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8652/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/720
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/720/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/720/comments
https://api.github.com/repos/ollama/ollama/issues/720/events
https://github.com/ollama/ollama/pull/720
1,930,612,519
PR_kwDOJ0Z1Ps5cId8b
720
relay model runner error message to client
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-10-06T16:52:04
2023-10-12T15:16:38
2023-10-12T15:16:37
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/720", "html_url": "https://github.com/ollama/ollama/pull/720", "diff_url": "https://github.com/ollama/ollama/pull/720.diff", "patch_url": "https://github.com/ollama/ollama/pull/720.patch", "merged_at": "2023-10-12T15:16:37" }
This got missed in the migration to subprocesses. Old error displayed in CLI: ``` failed to start llama runner ``` New error (will relay the actual error from the model runner): ``` Error: llama runner failed: out of memory ``` resolves #630
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/720/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/720/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2867
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2867/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2867/comments
https://api.github.com/repos/ollama/ollama/issues/2867/events
https://github.com/ollama/ollama/pull/2867
2,164,036,591
PR_kwDOJ0Z1Ps5ocY1E
2,867
Added ollama4j-web-ui to the list of community developed user interfaces
{ "login": "amithkoujalgi", "id": 1876165, "node_id": "MDQ6VXNlcjE4NzYxNjU=", "avatar_url": "https://avatars.githubusercontent.com/u/1876165?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amithkoujalgi", "html_url": "https://github.com/amithkoujalgi", "followers_url": "https://api.github.com/users/amithkoujalgi/followers", "following_url": "https://api.github.com/users/amithkoujalgi/following{/other_user}", "gists_url": "https://api.github.com/users/amithkoujalgi/gists{/gist_id}", "starred_url": "https://api.github.com/users/amithkoujalgi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amithkoujalgi/subscriptions", "organizations_url": "https://api.github.com/users/amithkoujalgi/orgs", "repos_url": "https://api.github.com/users/amithkoujalgi/repos", "events_url": "https://api.github.com/users/amithkoujalgi/events{/privacy}", "received_events_url": "https://api.github.com/users/amithkoujalgi/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-03-01T19:24:51
2024-04-28T14:34:04
2024-04-28T14:34:04
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2867", "html_url": "https://github.com/ollama/ollama/pull/2867", "diff_url": "https://github.com/ollama/ollama/pull/2867.diff", "patch_url": "https://github.com/ollama/ollama/pull/2867.patch", "merged_at": null }
Added ollama4j-web-ui to the list of community developed user interfaces.
{ "login": "amithkoujalgi", "id": 1876165, "node_id": "MDQ6VXNlcjE4NzYxNjU=", "avatar_url": "https://avatars.githubusercontent.com/u/1876165?v=4", "gravatar_id": "", "url": "https://api.github.com/users/amithkoujalgi", "html_url": "https://github.com/amithkoujalgi", "followers_url": "https://api.github.com/users/amithkoujalgi/followers", "following_url": "https://api.github.com/users/amithkoujalgi/following{/other_user}", "gists_url": "https://api.github.com/users/amithkoujalgi/gists{/gist_id}", "starred_url": "https://api.github.com/users/amithkoujalgi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/amithkoujalgi/subscriptions", "organizations_url": "https://api.github.com/users/amithkoujalgi/orgs", "repos_url": "https://api.github.com/users/amithkoujalgi/repos", "events_url": "https://api.github.com/users/amithkoujalgi/events{/privacy}", "received_events_url": "https://api.github.com/users/amithkoujalgi/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2867/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2867/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2060
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2060/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2060/comments
https://api.github.com/repos/ollama/ollama/issues/2060/events
https://github.com/ollama/ollama/pull/2060
2,089,279,052
PR_kwDOJ0Z1Ps5kemlc
2,060
fix show handler
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-18T23:37:30
2024-01-19T00:02:28
2024-01-19T00:02:28
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2060", "html_url": "https://github.com/ollama/ollama/pull/2060", "diff_url": "https://github.com/ollama/ollama/pull/2060.diff", "patch_url": "https://github.com/ollama/ollama/pull/2060.patch", "merged_at": "2024-01-19T00:02:28" }
the show handler is unique in the sense that the request struct is passed directly to `GetModel` while every other handler deconstructs the request into its parameters. therefore the pattern of setting a local variable `model` and setting it to req.Model or req.Name doesn't work
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2060/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2060/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4326
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4326/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4326/comments
https://api.github.com/repos/ollama/ollama/issues/4326/events
https://github.com/ollama/ollama/pull/4326
2,290,427,873
PR_kwDOJ0Z1Ps5vItCX
4,326
Integration fixes
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-10T21:20:32
2024-05-10T21:26:02
2024-05-10T21:25:59
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4326", "html_url": "https://github.com/ollama/ollama/pull/4326", "diff_url": "https://github.com/ollama/ollama/pull/4326.diff", "patch_url": "https://github.com/ollama/ollama/pull/4326.patch", "merged_at": "2024-05-10T21:25:59" }
null
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4326/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4326/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4038
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4038/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4038/comments
https://api.github.com/repos/ollama/ollama/issues/4038/events
https://github.com/ollama/ollama/issues/4038
2,270,321,683
I_kwDOJ0Z1Ps6HUlwT
4,038
ollama run codeqwen TLS handshake timeout
{ "login": "Alchemistqqqq", "id": 146717415, "node_id": "U_kgDOCL665w", "avatar_url": "https://avatars.githubusercontent.com/u/146717415?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Alchemistqqqq", "html_url": "https://github.com/Alchemistqqqq", "followers_url": "https://api.github.com/users/Alchemistqqqq/followers", "following_url": "https://api.github.com/users/Alchemistqqqq/following{/other_user}", "gists_url": "https://api.github.com/users/Alchemistqqqq/gists{/gist_id}", "starred_url": "https://api.github.com/users/Alchemistqqqq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Alchemistqqqq/subscriptions", "organizations_url": "https://api.github.com/users/Alchemistqqqq/orgs", "repos_url": "https://api.github.com/users/Alchemistqqqq/repos", "events_url": "https://api.github.com/users/Alchemistqqqq/events{/privacy}", "received_events_url": "https://api.github.com/users/Alchemistqqqq/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677370291, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw", "url": "https://api.github.com/repos/ollama/ollama/labels/networking", "name": "networking", "color": "0B5368", "default": false, "description": "Issues relating to ollama pull and push" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
4
2024-04-30T01:38:12
2024-05-09T21:05:56
2024-05-09T21:05:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? After I manually installed ollama locally, I started to try to run models, not only qianwne, but also llama3. But ping ollama.com does work. Since I am building ollama on the server, I was wondering if it would be possible to download the model file locally and drag it to the specified path on the server. If you can do this, could you give me some help? I don't know what path the pulled model should be placed under, thank you. ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.1.32
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4038/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4038/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2550
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2550/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2550/comments
https://api.github.com/repos/ollama/ollama/issues/2550/events
https://github.com/ollama/ollama/issues/2550
2,139,324,363
I_kwDOJ0Z1Ps5_g3_L
2,550
WIndows questions sorry
{ "login": "drewskidang", "id": 120294157, "node_id": "U_kgDOByuLDQ", "avatar_url": "https://avatars.githubusercontent.com/u/120294157?v=4", "gravatar_id": "", "url": "https://api.github.com/users/drewskidang", "html_url": "https://github.com/drewskidang", "followers_url": "https://api.github.com/users/drewskidang/followers", "following_url": "https://api.github.com/users/drewskidang/following{/other_user}", "gists_url": "https://api.github.com/users/drewskidang/gists{/gist_id}", "starred_url": "https://api.github.com/users/drewskidang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/drewskidang/subscriptions", "organizations_url": "https://api.github.com/users/drewskidang/orgs", "repos_url": "https://api.github.com/users/drewskidang/repos", "events_url": "https://api.github.com/users/drewskidang/events{/privacy}", "received_events_url": "https://api.github.com/users/drewskidang/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-02-16T19:49:52
2024-02-20T03:49:06
2024-02-20T03:49:06
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How do you login using windows since theres no cat funciton
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2550/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2550/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2213
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2213/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2213/comments
https://api.github.com/repos/ollama/ollama/issues/2213/events
https://github.com/ollama/ollama/issues/2213
2,102,755,684
I_kwDOJ0Z1Ps59VYFk
2,213
Interleaving text and images (for few-shot learning)
{ "login": "delenius", "id": 7936653, "node_id": "MDQ6VXNlcjc5MzY2NTM=", "avatar_url": "https://avatars.githubusercontent.com/u/7936653?v=4", "gravatar_id": "", "url": "https://api.github.com/users/delenius", "html_url": "https://github.com/delenius", "followers_url": "https://api.github.com/users/delenius/followers", "following_url": "https://api.github.com/users/delenius/following{/other_user}", "gists_url": "https://api.github.com/users/delenius/gists{/gist_id}", "starred_url": "https://api.github.com/users/delenius/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/delenius/subscriptions", "organizations_url": "https://api.github.com/users/delenius/orgs", "repos_url": "https://api.github.com/users/delenius/repos", "events_url": "https://api.github.com/users/delenius/events{/privacy}", "received_events_url": "https://api.github.com/users/delenius/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
3
2024-01-26T19:13:36
2024-09-04T18:31:10
2024-09-04T18:31:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It does not appear to be possible (e.g. with llava) to interleave images and text (or is it?). This would be necessary in order to give some few-shot examples of image-text pairs, and then a final image that we want to generate text for. For example, the [OpenAI API](https://platform.openai.com/docs/guides/vision) allows for this by having the `content` field be a list, where each entry can be either text, or a base64-encoded image. (The examples in their docs do not show it, but it is indeed possible to interleave images and text arbitrarily using that API.) I am not sure this is possible with the underlying llava model (or others), but if it is, it would be a great feature to have.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2213/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2213/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2972
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2972/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2972/comments
https://api.github.com/repos/ollama/ollama/issues/2972/events
https://github.com/ollama/ollama/issues/2972
2,173,059,774
I_kwDOJ0Z1Ps6BhkK-
2,972
How to run Windows version Ollama on AMD GPU?
{ "login": "TM119", "id": 154800639, "node_id": "U_kgDOCToR_w", "avatar_url": "https://avatars.githubusercontent.com/u/154800639?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TM119", "html_url": "https://github.com/TM119", "followers_url": "https://api.github.com/users/TM119/followers", "following_url": "https://api.github.com/users/TM119/following{/other_user}", "gists_url": "https://api.github.com/users/TM119/gists{/gist_id}", "starred_url": "https://api.github.com/users/TM119/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TM119/subscriptions", "organizations_url": "https://api.github.com/users/TM119/orgs", "repos_url": "https://api.github.com/users/TM119/repos", "events_url": "https://api.github.com/users/TM119/events{/privacy}", "received_events_url": "https://api.github.com/users/TM119/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-03-07T06:27:30
2024-03-07T17:57:31
2024-03-07T07:06:07
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have a W6800, apparently windows version Ollama is running models on CPU rather than GPU. Will AMD GPU be supported?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2972/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2972/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2275
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2275/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2275/comments
https://api.github.com/repos/ollama/ollama/issues/2275/events
https://github.com/ollama/ollama/pull/2275
2,108,068,005
PR_kwDOJ0Z1Ps5ldpLl
2,275
Update troubleshooting.md with troubleshooting details for containers.
{ "login": "ltomes", "id": 4184677, "node_id": "MDQ6VXNlcjQxODQ2Nzc=", "avatar_url": "https://avatars.githubusercontent.com/u/4184677?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ltomes", "html_url": "https://github.com/ltomes", "followers_url": "https://api.github.com/users/ltomes/followers", "following_url": "https://api.github.com/users/ltomes/following{/other_user}", "gists_url": "https://api.github.com/users/ltomes/gists{/gist_id}", "starred_url": "https://api.github.com/users/ltomes/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ltomes/subscriptions", "organizations_url": "https://api.github.com/users/ltomes/orgs", "repos_url": "https://api.github.com/users/ltomes/repos", "events_url": "https://api.github.com/users/ltomes/events{/privacy}", "received_events_url": "https://api.github.com/users/ltomes/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-30T14:58:14
2024-11-21T09:36:15
2024-11-21T09:36:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2275", "html_url": "https://github.com/ollama/ollama/pull/2275", "diff_url": "https://github.com/ollama/ollama/pull/2275.diff", "patch_url": "https://github.com/ollama/ollama/pull/2275.patch", "merged_at": null }
Update troubleshooting.md with troubleshooting details for containers.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2275/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2275/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6903
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6903/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6903/comments
https://api.github.com/repos/ollama/ollama/issues/6903/events
https://github.com/ollama/ollama/issues/6903
2,540,387,692
I_kwDOJ0Z1Ps6Xaz1s
6,903
nexusraven:13b-v2-q2_K EOF
{ "login": "shuther", "id": 33225407, "node_id": "MDQ6VXNlcjMzMjI1NDA3", "avatar_url": "https://avatars.githubusercontent.com/u/33225407?v=4", "gravatar_id": "", "url": "https://api.github.com/users/shuther", "html_url": "https://github.com/shuther", "followers_url": "https://api.github.com/users/shuther/followers", "following_url": "https://api.github.com/users/shuther/following{/other_user}", "gists_url": "https://api.github.com/users/shuther/gists{/gist_id}", "starred_url": "https://api.github.com/users/shuther/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/shuther/subscriptions", "organizations_url": "https://api.github.com/users/shuther/orgs", "repos_url": "https://api.github.com/users/shuther/repos", "events_url": "https://api.github.com/users/shuther/events{/privacy}", "received_events_url": "https://api.github.com/users/shuther/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
3
2024-09-21T16:18:09
2024-11-06T00:19:49
2024-11-06T00:19:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? either using the API or the CLI, I can't get nexusraven:13b-v2-q2_K to work (other models work fine) ``` ollama pull nexusraven:13b-v2-q2_K OLLAMA_DEBUG="1" ollama run nexusraven:13b-v2-q2_K Error: Post "http://127.0.0.1:11434/api/generate": EOF ``` ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.3.11
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6903/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6903/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2406
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2406/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2406/comments
https://api.github.com/repos/ollama/ollama/issues/2406/events
https://github.com/ollama/ollama/issues/2406
2,124,374,930
I_kwDOJ0Z1Ps5-n2OS
2,406
systemctl stop ollama.service times out
{ "login": "donuts-are-good", "id": 96031819, "node_id": "U_kgDOBblUSw", "avatar_url": "https://avatars.githubusercontent.com/u/96031819?v=4", "gravatar_id": "", "url": "https://api.github.com/users/donuts-are-good", "html_url": "https://github.com/donuts-are-good", "followers_url": "https://api.github.com/users/donuts-are-good/followers", "following_url": "https://api.github.com/users/donuts-are-good/following{/other_user}", "gists_url": "https://api.github.com/users/donuts-are-good/gists{/gist_id}", "starred_url": "https://api.github.com/users/donuts-are-good/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/donuts-are-good/subscriptions", "organizations_url": "https://api.github.com/users/donuts-are-good/orgs", "repos_url": "https://api.github.com/users/donuts-are-good/repos", "events_url": "https://api.github.com/users/donuts-are-good/events{/privacy}", "received_events_url": "https://api.github.com/users/donuts-are-good/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-08T05:13:52
2024-02-08T05:14:40
2024-02-08T05:14:40
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
As described, when I try to stop the systemd unit on Fedora linux from ollama (current version), I get this: Failed to stop ollama.service: Connection timed out
{ "login": "donuts-are-good", "id": 96031819, "node_id": "U_kgDOBblUSw", "avatar_url": "https://avatars.githubusercontent.com/u/96031819?v=4", "gravatar_id": "", "url": "https://api.github.com/users/donuts-are-good", "html_url": "https://github.com/donuts-are-good", "followers_url": "https://api.github.com/users/donuts-are-good/followers", "following_url": "https://api.github.com/users/donuts-are-good/following{/other_user}", "gists_url": "https://api.github.com/users/donuts-are-good/gists{/gist_id}", "starred_url": "https://api.github.com/users/donuts-are-good/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/donuts-are-good/subscriptions", "organizations_url": "https://api.github.com/users/donuts-are-good/orgs", "repos_url": "https://api.github.com/users/donuts-are-good/repos", "events_url": "https://api.github.com/users/donuts-are-good/events{/privacy}", "received_events_url": "https://api.github.com/users/donuts-are-good/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2406/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2406/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7055
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7055/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7055/comments
https://api.github.com/repos/ollama/ollama/issues/7055/events
https://github.com/ollama/ollama/pull/7055
2,558,182,298
PR_kwDOJ0Z1Ps59MxUY
7,055
Fixing generate failure due to global gpgsign setting
{ "login": "wikki01", "id": 150323920, "node_id": "U_kgDOCPXC0A", "avatar_url": "https://avatars.githubusercontent.com/u/150323920?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wikki01", "html_url": "https://github.com/wikki01", "followers_url": "https://api.github.com/users/wikki01/followers", "following_url": "https://api.github.com/users/wikki01/following{/other_user}", "gists_url": "https://api.github.com/users/wikki01/gists{/gist_id}", "starred_url": "https://api.github.com/users/wikki01/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wikki01/subscriptions", "organizations_url": "https://api.github.com/users/wikki01/orgs", "repos_url": "https://api.github.com/users/wikki01/repos", "events_url": "https://api.github.com/users/wikki01/events{/privacy}", "received_events_url": "https://api.github.com/users/wikki01/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-10-01T04:57:36
2024-11-21T19:26:33
2024-11-21T19:26:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7055", "html_url": "https://github.com/ollama/ollama/pull/7055", "diff_url": "https://github.com/ollama/ollama/pull/7055.diff", "patch_url": "https://github.com/ollama/ollama/pull/7055.patch", "merged_at": null }
Hi all, first time commit here so please let me know if I'm missing anything. Found a small issue in the generate script for `llama.cpp` where the script patches with a placeholder user. Since I have `commit.gpgsign` enabled globally on my system, this was causing the following error during `go generate ./...`. ``` $ go generate ./... <snip> + git submodule init + git submodule update --force ../llama.cpp Submodule path '../llama.cpp': checked out '8962422b1c6f9b8b15f5aeaea42600bcc2d44177' + apply_patches + for patch in ../patches/*.patch + git -c user.name=nobody -c 'user.email=<>' -C ../llama.cpp am ../patches/0000-cmakelist.patch Applying: patch cmakelist error: gpg failed to sign the data: gpg: skipped "nobody <>": No secret key [GNUPG:] INV_SGNR 9 nobody <> [GNUPG:] FAILURE sign 17 gpg: signing failed: No secret key fatal: failed to write commit object llm/generate/generate_linux.go:3: running "bash": exit status 128 ``` Since this happens in a submodule, setting `commit.gpgsign` to `false` for the top-level git repository does not have any effect. A workaround in the meantime is to disable global gpg signing temporarily, I.E. ``` git config --global commit.gpgsign false go generate ./... git config --global commit.gpgsign true ``` Cheers!
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7055/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7055/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2579
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2579/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2579/comments
https://api.github.com/repos/ollama/ollama/issues/2579/events
https://github.com/ollama/ollama/issues/2579
2,141,099,788
I_kwDOJ0Z1Ps5_npcM
2,579
Use pkg-config to find llama-cpp libs
{ "login": "happysalada", "id": 5317234, "node_id": "MDQ6VXNlcjUzMTcyMzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/5317234?v=4", "gravatar_id": "", "url": "https://api.github.com/users/happysalada", "html_url": "https://github.com/happysalada", "followers_url": "https://api.github.com/users/happysalada/followers", "following_url": "https://api.github.com/users/happysalada/following{/other_user}", "gists_url": "https://api.github.com/users/happysalada/gists{/gist_id}", "starred_url": "https://api.github.com/users/happysalada/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/happysalada/subscriptions", "organizations_url": "https://api.github.com/users/happysalada/orgs", "repos_url": "https://api.github.com/users/happysalada/repos", "events_url": "https://api.github.com/users/happysalada/events{/privacy}", "received_events_url": "https://api.github.com/users/happysalada/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-02-18T17:28:59
2024-10-23T17:53:17
2024-10-23T17:53:17
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi first of all, thank you for ollama, its amazing ! Im one of the person trying to maintain this on nixos. Since 0.1.19 i think llama-cpp is directly included as a submodule in the repo. Also the .so objects are directly referenced by path in the go build files. Would it be possible to use pkg-config to find those files ? This would make building ollama from source much easier as llama-cpp can just be specified as a dependency. I understand that you might have a million more important things to do, i just want to start the discussion.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2579/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2579/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5489
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5489/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5489/comments
https://api.github.com/repos/ollama/ollama/issues/5489/events
https://github.com/ollama/ollama/issues/5489
2,391,402,086
I_kwDOJ0Z1Ps6OieZm
5,489
Running Hugging Face Models
{ "login": "theainerd", "id": 15798640, "node_id": "MDQ6VXNlcjE1Nzk4NjQw", "avatar_url": "https://avatars.githubusercontent.com/u/15798640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/theainerd", "html_url": "https://github.com/theainerd", "followers_url": "https://api.github.com/users/theainerd/followers", "following_url": "https://api.github.com/users/theainerd/following{/other_user}", "gists_url": "https://api.github.com/users/theainerd/gists{/gist_id}", "starred_url": "https://api.github.com/users/theainerd/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/theainerd/subscriptions", "organizations_url": "https://api.github.com/users/theainerd/orgs", "repos_url": "https://api.github.com/users/theainerd/repos", "events_url": "https://api.github.com/users/theainerd/events{/privacy}", "received_events_url": "https://api.github.com/users/theainerd/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-07-04T19:54:54
2024-07-08T22:43:05
2024-07-08T22:43:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi team, Can we load Hugging Face models directly using ollama?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5489/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5489/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7565
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7565/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7565/comments
https://api.github.com/repos/ollama/ollama/issues/7565/events
https://github.com/ollama/ollama/issues/7565
2,642,583,722
I_kwDOJ0Z1Ps6dgqCq
7,565
Linux ollama 0.4.0, 0.4.2, 0.4.5, 0.5.0 custom compile for AMD ROCm fails missing ggml_rocm in go compile
{ "login": "ganakee", "id": 23140586, "node_id": "MDQ6VXNlcjIzMTQwNTg2", "avatar_url": "https://avatars.githubusercontent.com/u/23140586?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ganakee", "html_url": "https://github.com/ganakee", "followers_url": "https://api.github.com/users/ganakee/followers", "following_url": "https://api.github.com/users/ganakee/following{/other_user}", "gists_url": "https://api.github.com/users/ganakee/gists{/gist_id}", "starred_url": "https://api.github.com/users/ganakee/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ganakee/subscriptions", "organizations_url": "https://api.github.com/users/ganakee/orgs", "repos_url": "https://api.github.com/users/ganakee/repos", "events_url": "https://api.github.com/users/ganakee/events{/privacy}", "received_events_url": "https://api.github.com/users/ganakee/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 7700262114, "node_id": "LA_kwDOJ0Z1Ps8AAAAByvis4g", "url": "https://api.github.com/repos/ollama/ollama/labels/build", "name": "build", "color": "006b75", "default": false, "description": "Issues relating to building ollama from source" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
14
2024-11-08T01:22:32
2024-12-07T17:11:44
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Report date: 2024-11-07 During a custom compile of ollama 0.4.0 on Linux (POP OS 22.04) for AMD ROCm GPUs (AMD 6650 GPU), the initial compile works. However, when trying to execute the go compile, the compile fails after about two minutes citing exit code 1 and saying the error is unable to find ggml_rocm. ROCm 6.0 ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.4.0
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7565/reactions", "total_count": 3, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 2, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7565/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/896
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/896/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/896/comments
https://api.github.com/repos/ollama/ollama/issues/896/events
https://github.com/ollama/ollama/issues/896
1,960,138,018
I_kwDOJ0Z1Ps501VUi
896
Multline editing prior lines not possible
{ "login": "jacksongoode", "id": 54308792, "node_id": "MDQ6VXNlcjU0MzA4Nzky", "avatar_url": "https://avatars.githubusercontent.com/u/54308792?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jacksongoode", "html_url": "https://github.com/jacksongoode", "followers_url": "https://api.github.com/users/jacksongoode/followers", "following_url": "https://api.github.com/users/jacksongoode/following{/other_user}", "gists_url": "https://api.github.com/users/jacksongoode/gists{/gist_id}", "starred_url": "https://api.github.com/users/jacksongoode/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jacksongoode/subscriptions", "organizations_url": "https://api.github.com/users/jacksongoode/orgs", "repos_url": "https://api.github.com/users/jacksongoode/repos", "events_url": "https://api.github.com/users/jacksongoode/events{/privacy}", "received_events_url": "https://api.github.com/users/jacksongoode/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2023-10-24T21:52:06
2024-07-28T10:25:38
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I would like to edit the text pasted after the multiline delimiter editable such that I can go back on previous lines to edit.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/896/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/896/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6200
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6200/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6200/comments
https://api.github.com/repos/ollama/ollama/issues/6200/events
https://github.com/ollama/ollama/issues/6200
2,450,896,081
I_kwDOJ0Z1Ps6SFbTR
6,200
Support --mlock on the command line. Also there are undocumented model file parameters
{ "login": "sdmorrey", "id": 146312245, "node_id": "U_kgDOCLiMNQ", "avatar_url": "https://avatars.githubusercontent.com/u/146312245?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sdmorrey", "html_url": "https://github.com/sdmorrey", "followers_url": "https://api.github.com/users/sdmorrey/followers", "following_url": "https://api.github.com/users/sdmorrey/following{/other_user}", "gists_url": "https://api.github.com/users/sdmorrey/gists{/gist_id}", "starred_url": "https://api.github.com/users/sdmorrey/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sdmorrey/subscriptions", "organizations_url": "https://api.github.com/users/sdmorrey/orgs", "repos_url": "https://api.github.com/users/sdmorrey/repos", "events_url": "https://api.github.com/users/sdmorrey/events{/privacy}", "received_events_url": "https://api.github.com/users/sdmorrey/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2024-08-06T13:28:28
2024-08-06T13:31:54
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? The original issue was that my Mac was experiencing an exponential slow down. Upon investigating the issue I noticed my RAM usage was very, very low, but my swap was climbing at the rate of about 1GB per 1K of filled context. In the end I discovered the --mlock flag in llama.cpp. I was in discord asking for help setting it since the command line Ollama straight up rejects it. Eventually we discovered that this is controlled by the model file parameter use_mlock Except that appears to be undocumented. In the same space I found a few other undocumented parameters. So in a nutshell, it would lovely if you could add --mlock to the command line since I don't want to have to edit every model file. Also if you could update the docs so use_mlock and the other undocumented parameters are documented, that would be awesome. Thanks for your hard work and such a great product! ### OS macOS ### GPU Intel ### CPU Intel ### Ollama version 0.3.3
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6200/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6200/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/1694
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1694/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1694/comments
https://api.github.com/repos/ollama/ollama/issues/1694/events
https://github.com/ollama/ollama/issues/1694
2,054,996,485
I_kwDOJ0Z1Ps56fMIF
1,694
Windows build is broken.
{ "login": "vinjn", "id": 558657, "node_id": "MDQ6VXNlcjU1ODY1Nw==", "avatar_url": "https://avatars.githubusercontent.com/u/558657?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vinjn", "html_url": "https://github.com/vinjn", "followers_url": "https://api.github.com/users/vinjn/followers", "following_url": "https://api.github.com/users/vinjn/following{/other_user}", "gists_url": "https://api.github.com/users/vinjn/gists{/gist_id}", "starred_url": "https://api.github.com/users/vinjn/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vinjn/subscriptions", "organizations_url": "https://api.github.com/users/vinjn/orgs", "repos_url": "https://api.github.com/users/vinjn/repos", "events_url": "https://api.github.com/users/vinjn/events{/privacy}", "received_events_url": "https://api.github.com/users/vinjn/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2023-12-24T06:06:33
2024-01-06T03:34:21
2024-01-06T03:34:21
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
```bash go build . # github.com/jmorganca/ollama/llm llm\llm.go:83:17: undefined: gpu.GetGPUInfo llm\llm.go:89:9: undefined: nativeInit llm\llm.go:92:109: undefined: extServer llm\llm.go:94:15: undefined: newDynamicShimExtServer llm\llm.go:101:9: undefined: newDefaultExtServer llm\llama.go:211:24: undefined: libEmbed llm\llama.go:218:19: undefined: libEmbed ```
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1694/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1694/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6775
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6775/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6775/comments
https://api.github.com/repos/ollama/ollama/issues/6775/events
https://github.com/ollama/ollama/pull/6775
2,522,379,982
PR_kwDOJ0Z1Ps57TbYq
6,775
Notify systemd that ollama server is ready
{ "login": "JingWoo", "id": 21989093, "node_id": "MDQ6VXNlcjIxOTg5MDkz", "avatar_url": "https://avatars.githubusercontent.com/u/21989093?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JingWoo", "html_url": "https://github.com/JingWoo", "followers_url": "https://api.github.com/users/JingWoo/followers", "following_url": "https://api.github.com/users/JingWoo/following{/other_user}", "gists_url": "https://api.github.com/users/JingWoo/gists{/gist_id}", "starred_url": "https://api.github.com/users/JingWoo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JingWoo/subscriptions", "organizations_url": "https://api.github.com/users/JingWoo/orgs", "repos_url": "https://api.github.com/users/JingWoo/repos", "events_url": "https://api.github.com/users/JingWoo/events{/privacy}", "received_events_url": "https://api.github.com/users/JingWoo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-09-12T13:19:22
2024-11-12T00:46:20
2024-11-12T00:46:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6775", "html_url": "https://github.com/ollama/ollama/pull/6775", "diff_url": "https://github.com/ollama/ollama/pull/6775.diff", "patch_url": "https://github.com/ollama/ollama/pull/6775.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6775/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6775/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6067
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6067/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6067/comments
https://api.github.com/repos/ollama/ollama/issues/6067/events
https://github.com/ollama/ollama/issues/6067
2,436,666,151
I_kwDOJ0Z1Ps6RPJMn
6,067
change client app installation path on windows
{ "login": "Kozmosa", "id": 19748251, "node_id": "MDQ6VXNlcjE5NzQ4MjUx", "avatar_url": "https://avatars.githubusercontent.com/u/19748251?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Kozmosa", "html_url": "https://github.com/Kozmosa", "followers_url": "https://api.github.com/users/Kozmosa/followers", "following_url": "https://api.github.com/users/Kozmosa/following{/other_user}", "gists_url": "https://api.github.com/users/Kozmosa/gists{/gist_id}", "starred_url": "https://api.github.com/users/Kozmosa/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Kozmosa/subscriptions", "organizations_url": "https://api.github.com/users/Kozmosa/orgs", "repos_url": "https://api.github.com/users/Kozmosa/repos", "events_url": "https://api.github.com/users/Kozmosa/events{/privacy}", "received_events_url": "https://api.github.com/users/Kozmosa/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-07-30T00:55:24
2024-07-30T17:13:54
2024-07-30T17:13:54
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
At present, when installing ollama windows preview version, there is only a default installation path(C:/users/user/local/..), which is unchangeable Hopefully, there would be another option for installation path.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6067/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2896
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2896/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2896/comments
https://api.github.com/repos/ollama/ollama/issues/2896/events
https://github.com/ollama/ollama/issues/2896
2,165,426,841
I_kwDOJ0Z1Ps6BEcqZ
2,896
View Model Licenses on the Model Registry / Library Page
{ "login": "startakovsky", "id": 689723, "node_id": "MDQ6VXNlcjY4OTcyMw==", "avatar_url": "https://avatars.githubusercontent.com/u/689723?v=4", "gravatar_id": "", "url": "https://api.github.com/users/startakovsky", "html_url": "https://github.com/startakovsky", "followers_url": "https://api.github.com/users/startakovsky/followers", "following_url": "https://api.github.com/users/startakovsky/following{/other_user}", "gists_url": "https://api.github.com/users/startakovsky/gists{/gist_id}", "starred_url": "https://api.github.com/users/startakovsky/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/startakovsky/subscriptions", "organizations_url": "https://api.github.com/users/startakovsky/orgs", "repos_url": "https://api.github.com/users/startakovsky/repos", "events_url": "https://api.github.com/users/startakovsky/events{/privacy}", "received_events_url": "https://api.github.com/users/startakovsky/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-03-03T15:55:33
2024-03-04T08:05:16
2024-03-04T08:05:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi! I was just poking around at all the amazing models you have listed on your growing registry. From a quality-of-life perspective, would be great to see the licenses at a glance available at this page: https://ollama.com/library I see a similar issue as this had been closed referring to a post-download inspectable license: https://github.com/ollama/ollama/issues/1509 For various reasons this doesn't solve the quality-of-life issue, not because it's on the cli, but because it requires downloading the model to inspect the license. Thanks! Steven
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2896/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2896/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/620
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/620/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/620/comments
https://api.github.com/repos/ollama/ollama/issues/620/events
https://github.com/ollama/ollama/issues/620
1,915,026,173
I_kwDOJ0Z1Ps5yJPr9
620
is there any plan to provide GUI interface?
{ "login": "pengjinning", "id": 1288209, "node_id": "MDQ6VXNlcjEyODgyMDk=", "avatar_url": "https://avatars.githubusercontent.com/u/1288209?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pengjinning", "html_url": "https://github.com/pengjinning", "followers_url": "https://api.github.com/users/pengjinning/followers", "following_url": "https://api.github.com/users/pengjinning/following{/other_user}", "gists_url": "https://api.github.com/users/pengjinning/gists{/gist_id}", "starred_url": "https://api.github.com/users/pengjinning/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pengjinning/subscriptions", "organizations_url": "https://api.github.com/users/pengjinning/orgs", "repos_url": "https://api.github.com/users/pengjinning/repos", "events_url": "https://api.github.com/users/pengjinning/events{/privacy}", "received_events_url": "https://api.github.com/users/pengjinning/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2023-09-27T08:54:58
2023-09-27T08:56:34
2023-09-27T08:56:34
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
is there any plan to provide GUI interface for ollama?
{ "login": "pengjinning", "id": 1288209, "node_id": "MDQ6VXNlcjEyODgyMDk=", "avatar_url": "https://avatars.githubusercontent.com/u/1288209?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pengjinning", "html_url": "https://github.com/pengjinning", "followers_url": "https://api.github.com/users/pengjinning/followers", "following_url": "https://api.github.com/users/pengjinning/following{/other_user}", "gists_url": "https://api.github.com/users/pengjinning/gists{/gist_id}", "starred_url": "https://api.github.com/users/pengjinning/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pengjinning/subscriptions", "organizations_url": "https://api.github.com/users/pengjinning/orgs", "repos_url": "https://api.github.com/users/pengjinning/repos", "events_url": "https://api.github.com/users/pengjinning/events{/privacy}", "received_events_url": "https://api.github.com/users/pengjinning/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/620/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/620/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8155
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8155/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8155/comments
https://api.github.com/repos/ollama/ollama/issues/8155/events
https://github.com/ollama/ollama/issues/8155
2,747,548,191
I_kwDOJ0Z1Ps6jxEIf
8,155
Clarification: "format" vs "tools" behaviours
{ "login": "VMinB12", "id": 93780551, "node_id": "U_kgDOBZb6Rw", "avatar_url": "https://avatars.githubusercontent.com/u/93780551?v=4", "gravatar_id": "", "url": "https://api.github.com/users/VMinB12", "html_url": "https://github.com/VMinB12", "followers_url": "https://api.github.com/users/VMinB12/followers", "following_url": "https://api.github.com/users/VMinB12/following{/other_user}", "gists_url": "https://api.github.com/users/VMinB12/gists{/gist_id}", "starred_url": "https://api.github.com/users/VMinB12/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/VMinB12/subscriptions", "organizations_url": "https://api.github.com/users/VMinB12/orgs", "repos_url": "https://api.github.com/users/VMinB12/repos", "events_url": "https://api.github.com/users/VMinB12/events{/privacy}", "received_events_url": "https://api.github.com/users/VMinB12/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
{ "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "ParthSareen", "id": 29360864, "node_id": "MDQ6VXNlcjI5MzYwODY0", "avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ParthSareen", "html_url": "https://github.com/ParthSareen", "followers_url": "https://api.github.com/users/ParthSareen/followers", "following_url": "https://api.github.com/users/ParthSareen/following{/other_user}", "gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}", "starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions", "organizations_url": "https://api.github.com/users/ParthSareen/orgs", "repos_url": "https://api.github.com/users/ParthSareen/repos", "events_url": "https://api.github.com/users/ParthSareen/events{/privacy}", "received_events_url": "https://api.github.com/users/ParthSareen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2024-12-18T11:25:11
2024-12-20T21:48:59
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Structured output formatting was recently released. There are now 2 different ways of structuring your output: 1. Pass a schema to the "format" argument to the API. 2. Convert the schema to a tool and pass it to the "tools" argument to the API. I understand that semantically, for structured output, it would make more sense to use the "format" option, whereas for tool use, the "tools" argument is more sensible. My question is if there is any difference under the hood? Specifically, I have the following questions: - Does `tools` insert the tool into the prompt using the tool tokens in the model chat template? - Does "format" insert the tool into the prompt using the tool tokens in the model chat template? - Does "format" ensure that the schema is respected, or only that the output is "json", but not necessarily "json" that complies to the schema? Is the behaviour exactly the same for "tools", or is there a difference? ### OS Linux ### GPU _No response_ ### CPU _No response_ ### Ollama version 0.5.4
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8155/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8155/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8356
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8356/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8356/comments
https://api.github.com/repos/ollama/ollama/issues/8356/events
https://github.com/ollama/ollama/issues/8356
2,776,861,370
I_kwDOJ0Z1Ps6lg4q6
8,356
Allow context to be set from the command line.
{ "login": "iplayfast", "id": 751306, "node_id": "MDQ6VXNlcjc1MTMwNg==", "avatar_url": "https://avatars.githubusercontent.com/u/751306?v=4", "gravatar_id": "", "url": "https://api.github.com/users/iplayfast", "html_url": "https://github.com/iplayfast", "followers_url": "https://api.github.com/users/iplayfast/followers", "following_url": "https://api.github.com/users/iplayfast/following{/other_user}", "gists_url": "https://api.github.com/users/iplayfast/gists{/gist_id}", "starred_url": "https://api.github.com/users/iplayfast/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/iplayfast/subscriptions", "organizations_url": "https://api.github.com/users/iplayfast/orgs", "repos_url": "https://api.github.com/users/iplayfast/repos", "events_url": "https://api.github.com/users/iplayfast/events{/privacy}", "received_events_url": "https://api.github.com/users/iplayfast/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
5
2025-01-09T05:22:14
2025-01-27T18:24:04
2025-01-11T00:05:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have a shell script called wizard that is simply ``` more ~/bin/wizard #!/bin/bash ollama run llama3.2 ``` It's a very useful script but I would like it to be able to set the context to a larger amount something like ``` ollama run llama3.2 --num_ctx 4096 ``` Current docs: ```How can I specify the context window size? By default, Ollama uses a context window size of 2048 tokens. To change this when using ollama run, use /set parameter: /set parameter num_ctx 4096 When using the API, specify the num_ctx parameter: curl http://localhost:11434/api/generate -d '{ "model": "llama3.2", "prompt": "Why is the sky blue?", "options": { "num_ctx": 4096 } }' ````
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8356/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8356/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8429
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8429/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8429/comments
https://api.github.com/repos/ollama/ollama/issues/8429/events
https://github.com/ollama/ollama/issues/8429
2,788,385,663
I_kwDOJ0Z1Ps6mM2N_
8,429
How do we EXPORT ollama responses to file TXT/JSON ?
{ "login": "GAC-Machine", "id": 141077601, "node_id": "U_kgDOCGisYQ", "avatar_url": "https://avatars.githubusercontent.com/u/141077601?v=4", "gravatar_id": "", "url": "https://api.github.com/users/GAC-Machine", "html_url": "https://github.com/GAC-Machine", "followers_url": "https://api.github.com/users/GAC-Machine/followers", "following_url": "https://api.github.com/users/GAC-Machine/following{/other_user}", "gists_url": "https://api.github.com/users/GAC-Machine/gists{/gist_id}", "starred_url": "https://api.github.com/users/GAC-Machine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/GAC-Machine/subscriptions", "organizations_url": "https://api.github.com/users/GAC-Machine/orgs", "repos_url": "https://api.github.com/users/GAC-Machine/repos", "events_url": "https://api.github.com/users/GAC-Machine/events{/privacy}", "received_events_url": "https://api.github.com/users/GAC-Machine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
1
2025-01-14T22:18:54
2025-01-28T21:16:32
2025-01-28T21:16:32
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
OS : Windows 11 Pro Model: ollama 3.2 Want to create logregistry file (trivially a txt/JSON file). I want a JSON structured file like: ``` { "Model" : "name_of-model", "running environment" : "Win-MacOS-Linux", "conversation": [ "prompt1" : "my_input_text" "response1" : "ollama_output_text" }, ... //so on ] } ``` Alternatively it would be adequate a txt file like: ``` Model name_of_model Running environment Win_MacOS_Linux ConversationID Prompt-Input Response-Output 00000001 [my_input_text] [ollama_output_text] ``` Referencing this issue on Github https://github.com/ollama/ollama/issues/1710 dated #Dec, 25, 2023. The command provided is for Unix-Like system (that is Linux, MacOS). `ollama run llama3.2 "prompt" >> response.txt` What about Windows ? I tried it, but it doesn't work.
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8429/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8429/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5987
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5987/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5987/comments
https://api.github.com/repos/ollama/ollama/issues/5987/events
https://github.com/ollama/ollama/pull/5987
2,432,454,943
PR_kwDOJ0Z1Ps52l6sH
5,987
llm: keep patch for llama 3 rope factors
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-26T15:16:34
2024-07-26T22:20:54
2024-07-26T22:20:53
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5987", "html_url": "https://github.com/ollama/ollama/pull/5987", "diff_url": "https://github.com/ollama/ollama/pull/5987.diff", "patch_url": "https://github.com/ollama/ollama/pull/5987.patch", "merged_at": "2024-07-26T22:20:53" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5987/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5987/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8441
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8441/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8441/comments
https://api.github.com/repos/ollama/ollama/issues/8441/events
https://github.com/ollama/ollama/issues/8441
2,789,517,382
I_kwDOJ0Z1Ps6mRKhG
8,441
Segmentation fault
{ "login": "eviejw12", "id": 195107358, "node_id": "U_kgDOC6EaHg", "avatar_url": "https://avatars.githubusercontent.com/u/195107358?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eviejw12", "html_url": "https://github.com/eviejw12", "followers_url": "https://api.github.com/users/eviejw12/followers", "following_url": "https://api.github.com/users/eviejw12/following{/other_user}", "gists_url": "https://api.github.com/users/eviejw12/gists{/gist_id}", "starred_url": "https://api.github.com/users/eviejw12/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eviejw12/subscriptions", "organizations_url": "https://api.github.com/users/eviejw12/orgs", "repos_url": "https://api.github.com/users/eviejw12/repos", "events_url": "https://api.github.com/users/eviejw12/events{/privacy}", "received_events_url": "https://api.github.com/users/eviejw12/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2025-01-15T11:13:38
2025-01-28T21:17:41
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Dear Ollama Team, I encountered an issue while attempting to run the 'ollama service' command after installation. Here are the detailed steps I followed and the error message I received: 1. I downloaded the 'ollama-linux-amd64.tgz' file from your official source because my research network has some access limitations. 2. I unzipped the file to the path '/usr/local/bin/ollama'. 3. When I executed the command 'ollama service', I encountered a 'Segmentation fault' error. I would appreciate it if you could look into this issue and provide any possible solutions or workarounds. Thank you for your attention to this matter. Best regards ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8441/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8441/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/892
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/892/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/892/comments
https://api.github.com/repos/ollama/ollama/issues/892/events
https://github.com/ollama/ollama/issues/892
1,959,769,436
I_kwDOJ0Z1Ps50z7Vc
892
permissions error on `ollama create` on Linux
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg", "url": "https://api.github.com/repos/ollama/ollama/labels/linux", "name": "linux", "color": "516E70", "default": false, "description": "" } ]
closed
false
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2023-10-24T17:47:57
2023-11-16T00:41:14
2023-11-16T00:41:14
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
On Linux, Ollama runs as another `ollama` user. This causes permission errors on `ollama create`. A fix is in progress, tracked here: https://github.com/jmorganca/ollama/issues/891 In the meantime, current workaround is to add `ollama` to the user's group so Ollama can access group-accessible files: ``` usermod -a -G $(id -gn $USER) ollama ``` Then re-start the `ollama` service ``` sudo systemctl restart ollama ```
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/892/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/892/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4927
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4927/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4927/comments
https://api.github.com/repos/ollama/ollama/issues/4927/events
https://github.com/ollama/ollama/issues/4927
2,341,541,806
I_kwDOJ0Z1Ps6LkReu
4,927
Support for zpoint_large_embedding_zh
{ "login": "Leroy-X", "id": 13515498, "node_id": "MDQ6VXNlcjEzNTE1NDk4", "avatar_url": "https://avatars.githubusercontent.com/u/13515498?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Leroy-X", "html_url": "https://github.com/Leroy-X", "followers_url": "https://api.github.com/users/Leroy-X/followers", "following_url": "https://api.github.com/users/Leroy-X/following{/other_user}", "gists_url": "https://api.github.com/users/Leroy-X/gists{/gist_id}", "starred_url": "https://api.github.com/users/Leroy-X/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Leroy-X/subscriptions", "organizations_url": "https://api.github.com/users/Leroy-X/orgs", "repos_url": "https://api.github.com/users/Leroy-X/repos", "events_url": "https://api.github.com/users/Leroy-X/events{/privacy}", "received_events_url": "https://api.github.com/users/Leroy-X/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
0
2024-06-08T08:06:24
2024-06-08T08:06:24
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
[https://huggingface.co/iampanda/zpoint_large_embedding_zh](https://huggingface.co/iampanda/zpoint_large_embedding_zh) Thanks.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4927/reactions", "total_count": 4, "+1": 4, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4927/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3926
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3926/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3926/comments
https://api.github.com/repos/ollama/ollama/issues/3926/events
https://github.com/ollama/ollama/pull/3926
2,264,711,671
PR_kwDOJ0Z1Ps5tyTmk
3,926
Fix release CI
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-26T00:27:37
2024-04-26T00:42:35
2024-04-26T00:42:31
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3926", "html_url": "https://github.com/ollama/ollama/pull/3926", "diff_url": "https://github.com/ollama/ollama/pull/3926.diff", "patch_url": "https://github.com/ollama/ollama/pull/3926.patch", "merged_at": "2024-04-26T00:42:31" }
download-artifact path was being used incorrectly. It is where to extract the zip not the files in the zip to extract. Default is workspace dir which is what we want, so omit it
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3926/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3926/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4912
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4912/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4912/comments
https://api.github.com/repos/ollama/ollama/issues/4912/events
https://github.com/ollama/ollama/issues/4912
2,340,820,945
I_kwDOJ0Z1Ps6LhhfR
4,912
Error: llama runner process has terminated: signal: aborted (core dumped)
{ "login": "mikestut", "id": 88723510, "node_id": "MDQ6VXNlcjg4NzIzNTEw", "avatar_url": "https://avatars.githubusercontent.com/u/88723510?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mikestut", "html_url": "https://github.com/mikestut", "followers_url": "https://api.github.com/users/mikestut/followers", "following_url": "https://api.github.com/users/mikestut/following{/other_user}", "gists_url": "https://api.github.com/users/mikestut/gists{/gist_id}", "starred_url": "https://api.github.com/users/mikestut/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mikestut/subscriptions", "organizations_url": "https://api.github.com/users/mikestut/orgs", "repos_url": "https://api.github.com/users/mikestut/repos", "events_url": "https://api.github.com/users/mikestut/events{/privacy}", "received_events_url": "https://api.github.com/users/mikestut/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
31
2024-06-07T16:39:11
2024-08-26T09:59:17
2024-06-09T17:20:43
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? When I run the CLI ollama run qwen2:72b-instruct-q2_K then download the model and run the model。 rError: llama runner process has terminated: signal: aborted (core dumped) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.38
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4912/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4912/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3153
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3153/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3153/comments
https://api.github.com/repos/ollama/ollama/issues/3153/events
https://github.com/ollama/ollama/pull/3153
2,187,248,013
PR_kwDOJ0Z1Ps5prcOi
3,153
Exporting Ollama to WebAssembly
{ "login": "samyfodil", "id": 76626119, "node_id": "MDQ6VXNlcjc2NjI2MTE5", "avatar_url": "https://avatars.githubusercontent.com/u/76626119?v=4", "gravatar_id": "", "url": "https://api.github.com/users/samyfodil", "html_url": "https://github.com/samyfodil", "followers_url": "https://api.github.com/users/samyfodil/followers", "following_url": "https://api.github.com/users/samyfodil/following{/other_user}", "gists_url": "https://api.github.com/users/samyfodil/gists{/gist_id}", "starred_url": "https://api.github.com/users/samyfodil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/samyfodil/subscriptions", "organizations_url": "https://api.github.com/users/samyfodil/orgs", "repos_url": "https://api.github.com/users/samyfodil/repos", "events_url": "https://api.github.com/users/samyfodil/events{/privacy}", "received_events_url": "https://api.github.com/users/samyfodil/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-03-14T20:51:46
2024-05-04T08:58:06
2024-03-14T20:54:05
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3153", "html_url": "https://github.com/ollama/ollama/pull/3153", "diff_url": "https://github.com/ollama/ollama/pull/3153.diff", "patch_url": "https://github.com/ollama/ollama/pull/3153.patch", "merged_at": null }
This PR exports ollama as a WebAssembly Host Module.
{ "login": "samyfodil", "id": 76626119, "node_id": "MDQ6VXNlcjc2NjI2MTE5", "avatar_url": "https://avatars.githubusercontent.com/u/76626119?v=4", "gravatar_id": "", "url": "https://api.github.com/users/samyfodil", "html_url": "https://github.com/samyfodil", "followers_url": "https://api.github.com/users/samyfodil/followers", "following_url": "https://api.github.com/users/samyfodil/following{/other_user}", "gists_url": "https://api.github.com/users/samyfodil/gists{/gist_id}", "starred_url": "https://api.github.com/users/samyfodil/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/samyfodil/subscriptions", "organizations_url": "https://api.github.com/users/samyfodil/orgs", "repos_url": "https://api.github.com/users/samyfodil/repos", "events_url": "https://api.github.com/users/samyfodil/events{/privacy}", "received_events_url": "https://api.github.com/users/samyfodil/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3153/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3153/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1225
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1225/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1225/comments
https://api.github.com/repos/ollama/ollama/issues/1225/events
https://github.com/ollama/ollama/pull/1225
2,004,921,810
PR_kwDOJ0Z1Ps5gDjSx
1,225
add Llama Coder
{ "login": "ex3ndr", "id": 400659, "node_id": "MDQ6VXNlcjQwMDY1OQ==", "avatar_url": "https://avatars.githubusercontent.com/u/400659?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ex3ndr", "html_url": "https://github.com/ex3ndr", "followers_url": "https://api.github.com/users/ex3ndr/followers", "following_url": "https://api.github.com/users/ex3ndr/following{/other_user}", "gists_url": "https://api.github.com/users/ex3ndr/gists{/gist_id}", "starred_url": "https://api.github.com/users/ex3ndr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ex3ndr/subscriptions", "organizations_url": "https://api.github.com/users/ex3ndr/orgs", "repos_url": "https://api.github.com/users/ex3ndr/repos", "events_url": "https://api.github.com/users/ex3ndr/events{/privacy}", "received_events_url": "https://api.github.com/users/ex3ndr/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-11-21T18:39:02
2023-11-21T19:08:20
2023-11-21T19:08:20
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1225", "html_url": "https://github.com/ollama/ollama/pull/1225", "diff_url": "https://github.com/ollama/ollama/pull/1225.diff", "patch_url": "https://github.com/ollama/ollama/pull/1225.patch", "merged_at": "2023-11-21T19:08:20" }
Hey, i have created a plugin for VS Code that integrates with ollama!
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1225/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1225/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1237
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1237/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1237/comments
https://api.github.com/repos/ollama/ollama/issues/1237/events
https://github.com/ollama/ollama/issues/1237
2,005,644,123
I_kwDOJ0Z1Ps53i7Nb
1,237
GPTQ / ExLlamaV2 (EXL2) quantisation
{ "login": "0xdevalias", "id": 753891, "node_id": "MDQ6VXNlcjc1Mzg5MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/753891?v=4", "gravatar_id": "", "url": "https://api.github.com/users/0xdevalias", "html_url": "https://github.com/0xdevalias", "followers_url": "https://api.github.com/users/0xdevalias/followers", "following_url": "https://api.github.com/users/0xdevalias/following{/other_user}", "gists_url": "https://api.github.com/users/0xdevalias/gists{/gist_id}", "starred_url": "https://api.github.com/users/0xdevalias/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/0xdevalias/subscriptions", "organizations_url": "https://api.github.com/users/0xdevalias/orgs", "repos_url": "https://api.github.com/users/0xdevalias/repos", "events_url": "https://api.github.com/users/0xdevalias/events{/privacy}", "received_events_url": "https://api.github.com/users/0xdevalias/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
6
2023-11-22T06:12:08
2024-08-15T14:24:39
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
# Feature Description Please provide a detailed written description of what you were trying to do, and what you expected `llama.cpp` to do as an enhancement. # Motivation It sounds like it's a fast/useful quantisation method: - https://towardsdatascience.com/exllamav2-the-fastest-library-to-run-llms-32aeda294d26 - https://github.com/mlabonne/llm-course/blob/main/Quantize_models_with_ExLlamaV2.ipynb - https://towardsdatascience.com/4-bit-quantization-with-gptq-36b0f4f02c34 - https://huggingface.co/blog/gptq-integration - https://oobabooga.github.io/blog/posts/gptq-awq-exl2-llamacpp/ - > A detailed comparison between GPTQ, AWQ, EXL2, q4_K_M, q4_K_S, and load_in_4bit: perplexity, VRAM, speed, model size, and loading time. ## See Also - https://github.com/ggerganov/llama.cpp/issues/4165
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1237/reactions", "total_count": 22, "+1": 22, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1237/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7234
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7234/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7234/comments
https://api.github.com/repos/ollama/ollama/issues/7234/events
https://github.com/ollama/ollama/issues/7234
2,593,744,618
I_kwDOJ0Z1Ps6amWbq
7,234
Add nemotron-4-340b-instruct
{ "login": "gileneusz", "id": 34601970, "node_id": "MDQ6VXNlcjM0NjAxOTcw", "avatar_url": "https://avatars.githubusercontent.com/u/34601970?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gileneusz", "html_url": "https://github.com/gileneusz", "followers_url": "https://api.github.com/users/gileneusz/followers", "following_url": "https://api.github.com/users/gileneusz/following{/other_user}", "gists_url": "https://api.github.com/users/gileneusz/gists{/gist_id}", "starred_url": "https://api.github.com/users/gileneusz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gileneusz/subscriptions", "organizations_url": "https://api.github.com/users/gileneusz/orgs", "repos_url": "https://api.github.com/users/gileneusz/repos", "events_url": "https://api.github.com/users/gileneusz/events{/privacy}", "received_events_url": "https://api.github.com/users/gileneusz/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2024-10-17T06:25:48
2024-11-25T13:36:12
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Since we recently discovered how good LLAMA 3.1 70B Nemotron is, maybe we could consider adding the 340B model... for those with abundant GPU resources - are there any?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7234/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7234/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8597
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8597/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8597/comments
https://api.github.com/repos/ollama/ollama/issues/8597/events
https://github.com/ollama/ollama/issues/8597
2,811,701,408
I_kwDOJ0Z1Ps6nlyig
8,597
Error: llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer (4x L40S, 384GB system RAM, Deepseek-R1)
{ "login": "orlyandico", "id": 1325420, "node_id": "MDQ6VXNlcjEzMjU0MjA=", "avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4", "gravatar_id": "", "url": "https://api.github.com/users/orlyandico", "html_url": "https://github.com/orlyandico", "followers_url": "https://api.github.com/users/orlyandico/followers", "following_url": "https://api.github.com/users/orlyandico/following{/other_user}", "gists_url": "https://api.github.com/users/orlyandico/gists{/gist_id}", "starred_url": "https://api.github.com/users/orlyandico/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/orlyandico/subscriptions", "organizations_url": "https://api.github.com/users/orlyandico/orgs", "repos_url": "https://api.github.com/users/orlyandico/repos", "events_url": "https://api.github.com/users/orlyandico/events{/privacy}", "received_events_url": "https://api.github.com/users/orlyandico/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
17
2025-01-26T17:56:11
2025-01-30T07:47:39
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am aware that 4 x L40S only has 192GB of VRAM, but the machine also has 384GB of system RAM. The error seems to indicate that 45108 MiB of RAM is being allocated with cudaMalloc and this is failing. This is very close to the GPU limit (46068 MiB). On my home setup (2x P40, admittedly not trying such a huge model) the GPU never gets close to its VRAM limit (typically only 22GB out of 24GB). ``` Jan 26 17:48:20 ip-172-31-3-6 ollama[1418]: llm_load_print_meta: rope_yarn_log_mul = 0.1000 Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: ggml_backend_cuda_buffer_type_alloc_buffer: allocating 45108.64 MiB on device 0: cudaMalloc failed: out of memory Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: llama_model_load: error loading model: unable to allocate CUDA0 buffer Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: llama_load_model_from_file: failed to load model Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: panic: unable to load model: /usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: goroutine 34 [running]: Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner.(*Server).loadModel(0xc0001a0000, {0x1b, 0x0, 0x0, 0x0, {0xc000194090, 0x4, 0x4}, 0xc00018a060, 0x0}, ...) Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner/runner.go:852 +0x3ad Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: created by github.com/ollama/ollama/llama/runner.Execute in goroutine 1 Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner/runner.go:970 +0xd0d Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:24.077Z level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server error" Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:24.578Z level=ERROR source=sched.go:455 msg="error loading llama server" error="llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer\nllama_load_model_from_file: failed to load model" Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: [GIN] 2025/01/26 - 17:51:24 | 500 | 3m6s | 127.0.0.1 | POST "/api/generate" Jan 26 17:51:30 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:30.190Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=5.6116838510000004 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1> Jan 26 17:51:30 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:30.947Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=6.369278328 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 26 17:51:31 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:31.707Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=7.128851916 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9 Jan 26 17:51:40 ip-172-31-3-6 ollama[1418]: [GIN] 2025/01/26 - 17:51:40 | 200 | 65.972µs | 127.0.0.1 | GET "/api/version" ubuntu@ip-172-31-3-6:~$ ollama --version ollama version is 0.5.7 ubuntu@ip-172-31-3-6:~$ free total used free shared buff/cache available Mem: 390837004 4152012 385439624 3248 1245368 383763992 Swap: 0 0 0 ubuntu@ip-172-31-3-6:~$ nvidia-smi Sun Jan 26 17:48:08 2025 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.144.03 Driver Version: 550.144.03 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA L40S On | 00000000:38:00.0 Off | 0 | | N/A 22C P8 22W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA L40S On | 00000000:3A:00.0 Off | 0 | | N/A 22C P8 21W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 2 NVIDIA L40S On | 00000000:3C:00.0 Off | 0 | | N/A 22C P8 44W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 3 NVIDIA L40S On | 00000000:3E:00.0 Off | 0 | | N/A 23C P8 22W / 350W | 4MiB / 46068MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ ``` ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8597/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8597/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6368
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6368/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6368/comments
https://api.github.com/repos/ollama/ollama/issues/6368/events
https://github.com/ollama/ollama/issues/6368
2,467,292,320
I_kwDOJ0Z1Ps6TD-Sg
6,368
Please add socks5 proxy func to download models !Because the internet in China is not accessible
{ "login": "icetech233", "id": 17383321, "node_id": "MDQ6VXNlcjE3MzgzMzIx", "avatar_url": "https://avatars.githubusercontent.com/u/17383321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/icetech233", "html_url": "https://github.com/icetech233", "followers_url": "https://api.github.com/users/icetech233/followers", "following_url": "https://api.github.com/users/icetech233/following{/other_user}", "gists_url": "https://api.github.com/users/icetech233/gists{/gist_id}", "starred_url": "https://api.github.com/users/icetech233/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/icetech233/subscriptions", "organizations_url": "https://api.github.com/users/icetech233/orgs", "repos_url": "https://api.github.com/users/icetech233/repos", "events_url": "https://api.github.com/users/icetech233/events{/privacy}", "received_events_url": "https://api.github.com/users/icetech233/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-08-15T03:17:01
2024-08-21T03:46:46
2024-08-15T20:19:46
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Please add socks5 proxy func to download models !Because the internet in China is not accessible
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6368/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6368/timeline
null
not_planned
false
https://api.github.com/repos/ollama/ollama/issues/1830
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1830/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1830/comments
https://api.github.com/repos/ollama/ollama/issues/1830/events
https://github.com/ollama/ollama/pull/1830
2,068,840,836
PR_kwDOJ0Z1Ps5jZCzz
1,830
Update README.md - Community Integrations - Ollama for Ruby
{ "login": "gbaptista", "id": 29520, "node_id": "MDQ6VXNlcjI5NTIw", "avatar_url": "https://avatars.githubusercontent.com/u/29520?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gbaptista", "html_url": "https://github.com/gbaptista", "followers_url": "https://api.github.com/users/gbaptista/followers", "following_url": "https://api.github.com/users/gbaptista/following{/other_user}", "gists_url": "https://api.github.com/users/gbaptista/gists{/gist_id}", "starred_url": "https://api.github.com/users/gbaptista/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gbaptista/subscriptions", "organizations_url": "https://api.github.com/users/gbaptista/orgs", "repos_url": "https://api.github.com/users/gbaptista/repos", "events_url": "https://api.github.com/users/gbaptista/events{/privacy}", "received_events_url": "https://api.github.com/users/gbaptista/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-06T21:26:22
2024-01-07T03:31:40
2024-01-07T03:31:39
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1830", "html_url": "https://github.com/ollama/ollama/pull/1830", "diff_url": "https://github.com/ollama/ollama/pull/1830.diff", "patch_url": "https://github.com/ollama/ollama/pull/1830.patch", "merged_at": "2024-01-07T03:31:39" }
Hi! Adding a new library for the Ruby language: - [Ollama for Ruby](https://github.com/gbaptista/ollama-ai) - [Ruby Gem](https://rubygems.org/gems/ollama-ai)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1830/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1830/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6976
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6976/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6976/comments
https://api.github.com/repos/ollama/ollama/issues/6976/events
https://github.com/ollama/ollama/issues/6976
2,549,980,410
I_kwDOJ0Z1Ps6X_Zz6
6,976
Support model allenai/OLMoE-1B-7B-0924
{ "login": "gigascake", "id": 36724511, "node_id": "MDQ6VXNlcjM2NzI0NTEx", "avatar_url": "https://avatars.githubusercontent.com/u/36724511?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gigascake", "html_url": "https://github.com/gigascake", "followers_url": "https://api.github.com/users/gigascake/followers", "following_url": "https://api.github.com/users/gigascake/following{/other_user}", "gists_url": "https://api.github.com/users/gigascake/gists{/gist_id}", "starred_url": "https://api.github.com/users/gigascake/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gigascake/subscriptions", "organizations_url": "https://api.github.com/users/gigascake/orgs", "repos_url": "https://api.github.com/users/gigascake/repos", "events_url": "https://api.github.com/users/gigascake/events{/privacy}", "received_events_url": "https://api.github.com/users/gigascake/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
1
2024-09-26T09:06:30
2024-10-03T15:12:45
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
i want check performance this model(MoE)
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6976/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6976/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2848
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2848/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2848/comments
https://api.github.com/repos/ollama/ollama/issues/2848/events
https://github.com/ollama/ollama/pull/2848
2,162,384,101
PR_kwDOJ0Z1Ps5oWsKe
2,848
Fix embeddings load model behavior
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-01T01:26:51
2024-03-01T01:40:57
2024-03-01T01:40:56
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2848", "html_url": "https://github.com/ollama/ollama/pull/2848", "diff_url": "https://github.com/ollama/ollama/pull/2848.diff", "patch_url": "https://github.com/ollama/ollama/pull/2848.patch", "merged_at": "2024-03-01T01:40:56" }
Fixes https://github.com/ollama/ollama/issues/2810. Also cleans up the `embedding_only` API option which had no effect.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2848/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2848/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2437
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2437/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2437/comments
https://api.github.com/repos/ollama/ollama/issues/2437/events
https://github.com/ollama/ollama/pull/2437
2,128,020,347
PR_kwDOJ0Z1Ps5mhy_l
2,437
replace strings buffer with hasher
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-02-10T00:54:52
2024-02-21T00:07:51
2024-02-21T00:07:50
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2437", "html_url": "https://github.com/ollama/ollama/pull/2437", "diff_url": "https://github.com/ollama/ollama/pull/2437.diff", "patch_url": "https://github.com/ollama/ollama/pull/2437.patch", "merged_at": "2024-02-21T00:07:50" }
the buffered value is going into the hasher eventually so write directly to the hasher instead
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2437/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2437/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/423
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/423/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/423/comments
https://api.github.com/repos/ollama/ollama/issues/423/events
https://github.com/ollama/ollama/issues/423
1,868,164,619
I_kwDOJ0Z1Ps5vWe4L
423
llamma.cpp breaking change deprecating GGML in favor of GGUF
{ "login": "jkleckner", "id": 1223485, "node_id": "MDQ6VXNlcjEyMjM0ODU=", "avatar_url": "https://avatars.githubusercontent.com/u/1223485?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jkleckner", "html_url": "https://github.com/jkleckner", "followers_url": "https://api.github.com/users/jkleckner/followers", "following_url": "https://api.github.com/users/jkleckner/following{/other_user}", "gists_url": "https://api.github.com/users/jkleckner/gists{/gist_id}", "starred_url": "https://api.github.com/users/jkleckner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jkleckner/subscriptions", "organizations_url": "https://api.github.com/users/jkleckner/orgs", "repos_url": "https://api.github.com/users/jkleckner/repos", "events_url": "https://api.github.com/users/jkleckner/events{/privacy}", "received_events_url": "https://api.github.com/users/jkleckner/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-08-26T16:27:32
2023-08-26T17:31:07
2023-08-26T17:31:06
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Note that the upstream llama.cpp project has now completely deprecated GGML in favor of GGUF [1]. How should the repository and user models adapt to this? [1] https://github.com/ggerganov/llama.cpp/pull/2398
{ "login": "jkleckner", "id": 1223485, "node_id": "MDQ6VXNlcjEyMjM0ODU=", "avatar_url": "https://avatars.githubusercontent.com/u/1223485?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jkleckner", "html_url": "https://github.com/jkleckner", "followers_url": "https://api.github.com/users/jkleckner/followers", "following_url": "https://api.github.com/users/jkleckner/following{/other_user}", "gists_url": "https://api.github.com/users/jkleckner/gists{/gist_id}", "starred_url": "https://api.github.com/users/jkleckner/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jkleckner/subscriptions", "organizations_url": "https://api.github.com/users/jkleckner/orgs", "repos_url": "https://api.github.com/users/jkleckner/repos", "events_url": "https://api.github.com/users/jkleckner/events{/privacy}", "received_events_url": "https://api.github.com/users/jkleckner/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/423/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/423/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5258
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5258/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5258/comments
https://api.github.com/repos/ollama/ollama/issues/5258/events
https://github.com/ollama/ollama/pull/5258
2,370,794,795
PR_kwDOJ0Z1Ps5zZbYA
5,258
Main
{ "login": "rpreslar4765", "id": 89657947, "node_id": "MDQ6VXNlcjg5NjU3OTQ3", "avatar_url": "https://avatars.githubusercontent.com/u/89657947?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rpreslar4765", "html_url": "https://github.com/rpreslar4765", "followers_url": "https://api.github.com/users/rpreslar4765/followers", "following_url": "https://api.github.com/users/rpreslar4765/following{/other_user}", "gists_url": "https://api.github.com/users/rpreslar4765/gists{/gist_id}", "starred_url": "https://api.github.com/users/rpreslar4765/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rpreslar4765/subscriptions", "organizations_url": "https://api.github.com/users/rpreslar4765/orgs", "repos_url": "https://api.github.com/users/rpreslar4765/repos", "events_url": "https://api.github.com/users/rpreslar4765/events{/privacy}", "received_events_url": "https://api.github.com/users/rpreslar4765/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-06-24T17:55:20
2024-06-24T17:59:55
2024-06-24T17:59:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5258", "html_url": "https://github.com/ollama/ollama/pull/5258", "diff_url": "https://github.com/ollama/ollama/pull/5258.diff", "patch_url": "https://github.com/ollama/ollama/pull/5258.patch", "merged_at": null }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5258/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5258/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2293
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2293/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2293/comments
https://api.github.com/repos/ollama/ollama/issues/2293/events
https://github.com/ollama/ollama/issues/2293
2,111,093,615
I_kwDOJ0Z1Ps591Ltv
2,293
Error: unmarshal
{ "login": "hemanth", "id": 18315, "node_id": "MDQ6VXNlcjE4MzE1", "avatar_url": "https://avatars.githubusercontent.com/u/18315?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hemanth", "html_url": "https://github.com/hemanth", "followers_url": "https://api.github.com/users/hemanth/followers", "following_url": "https://api.github.com/users/hemanth/following{/other_user}", "gists_url": "https://api.github.com/users/hemanth/gists{/gist_id}", "starred_url": "https://api.github.com/users/hemanth/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hemanth/subscriptions", "organizations_url": "https://api.github.com/users/hemanth/orgs", "repos_url": "https://api.github.com/users/hemanth/repos", "events_url": "https://api.github.com/users/hemanth/events{/privacy}", "received_events_url": "https://api.github.com/users/hemanth/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
10
2024-01-31T22:30:49
2024-05-02T01:24:23
2024-03-11T23:14:09
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Noticing a `Error: unmarshal: invalid character 'p' after top-level value` on `ollama run llava` <img width="923" alt="image" src="https://github.com/ollama/ollama/assets/18315/84f04d3e-a64e-4da5-9131-20ccf56cdeca"> `client version is 0.1.22`
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2293/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2293/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4088
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4088/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4088/comments
https://api.github.com/repos/ollama/ollama/issues/4088/events
https://github.com/ollama/ollama/issues/4088
2,274,072,575
I_kwDOJ0Z1Ps6Hi5f_
4,088
Colons in hostname cause an error on Windows
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
open
false
{ "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "bmizerany", "id": 46, "node_id": "MDQ6VXNlcjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/46?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bmizerany", "html_url": "https://github.com/bmizerany", "followers_url": "https://api.github.com/users/bmizerany/followers", "following_url": "https://api.github.com/users/bmizerany/following{/other_user}", "gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}", "starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions", "organizations_url": "https://api.github.com/users/bmizerany/orgs", "repos_url": "https://api.github.com/users/bmizerany/repos", "events_url": "https://api.github.com/users/bmizerany/events{/privacy}", "received_events_url": "https://api.github.com/users/bmizerany/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-05-01T19:36:20
2024-05-02T17:53:48
null
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Including a colon in the hostname such as `localhost:6000/my/model` causes an error on windows: ``` $ ollama create -f ../models/Modelfile localhost:3000/library/tmp transferring model data reading model metadata creating template layer using already created layer sha256:4fed7364ee3e0c7cb4fe0880148bfdfcd1b630981efa0802a6b62ee52e7da97e using already created layer sha256:fa8235e5b48faca34e3ca98cf4f694ef08bd216d28b58071a1f85b1d50cb814d using already created layer sha256:d47ab88b61ba20ed39a1b205a7d5a8e201dcf09107e6b05f128778c32baa4a69 using already created layer sha256:e14c21df1d1c6baff291558d0da327e5aabe4d4e7dd97bc1bf173e535f2ab210 using already created layer sha256:cb4e88e025cabd86f873cd94972dc82cfd883336b2294882a5c28af996a4db92 writing manifest Error: mkdir C:\Users\jmorg\.ollama\models\manifests\localhost:3000: The directory name is invalid. ``` ### OS Windows ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4088/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4088/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/212
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/212/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/212/comments
https://api.github.com/repos/ollama/ollama/issues/212/events
https://github.com/ollama/ollama/pull/212
1,820,946,267
PR_kwDOJ0Z1Ps5WXbPk
212
fix multiline string
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-07-25T18:52:00
2023-07-25T18:53:55
2023-07-25T18:53:51
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/212", "html_url": "https://github.com/ollama/ollama/pull/212", "diff_url": "https://github.com/ollama/ollama/pull/212.diff", "patch_url": "https://github.com/ollama/ollama/pull/212.patch", "merged_at": "2023-07-25T18:53:51" }
the data needs to remove the multiline quotes but include the command: e.g. ``` TEMPLATE """ my template values """ ``` should be ``` TEMPLATE my template values ``` after scanning
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/212/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/212/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7018
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7018/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7018/comments
https://api.github.com/repos/ollama/ollama/issues/7018/events
https://github.com/ollama/ollama/issues/7018
2,554,099,512
I_kwDOJ0Z1Ps6YPHc4
7,018
AMD Radeon Pro 5700 XT 16GB GPU Support
{ "login": "tomasznazarenko", "id": 28782643, "node_id": "MDQ6VXNlcjI4NzgyNjQz", "avatar_url": "https://avatars.githubusercontent.com/u/28782643?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tomasznazarenko", "html_url": "https://github.com/tomasznazarenko", "followers_url": "https://api.github.com/users/tomasznazarenko/followers", "following_url": "https://api.github.com/users/tomasznazarenko/following{/other_user}", "gists_url": "https://api.github.com/users/tomasznazarenko/gists{/gist_id}", "starred_url": "https://api.github.com/users/tomasznazarenko/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tomasznazarenko/subscriptions", "organizations_url": "https://api.github.com/users/tomasznazarenko/orgs", "repos_url": "https://api.github.com/users/tomasznazarenko/repos", "events_url": "https://api.github.com/users/tomasznazarenko/events{/privacy}", "received_events_url": "https://api.github.com/users/tomasznazarenko/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-09-28T08:04:16
2024-09-28T22:44:04
2024-09-28T22:44:04
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi, Would it be possible to add support for AMD Radeon Pro 5700 XT 16GB VRAM GPU? System: macOS Sequoia CPU: 3,8 GHz 8-Core Intel Core i7 RAM: 128GB Currently when using such hardware Ollama utilizes only CPU. The system is from 2020, but it is still a capable machine. Best,
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7018/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7018/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7592
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7592/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7592/comments
https://api.github.com/repos/ollama/ollama/issues/7592/events
https://github.com/ollama/ollama/issues/7592
2,646,662,886
I_kwDOJ0Z1Ps6dwN7m
7,592
Wrong Prompt Token Report from Ignoring Image Token Count
{ "login": "chigkim", "id": 22120994, "node_id": "MDQ6VXNlcjIyMTIwOTk0", "avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chigkim", "html_url": "https://github.com/chigkim", "followers_url": "https://api.github.com/users/chigkim/followers", "following_url": "https://api.github.com/users/chigkim/following{/other_user}", "gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}", "starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chigkim/subscriptions", "organizations_url": "https://api.github.com/users/chigkim/orgs", "repos_url": "https://api.github.com/users/chigkim/repos", "events_url": "https://api.github.com/users/chigkim/events{/privacy}", "received_events_url": "https://api.github.com/users/chigkim/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-11-09T23:31:37
2024-11-11T18:47:55
2024-11-11T18:47:54
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? If you run llama3.2-vision with /set verbose, it looks like it only counts tokens for text ignoring tokens from image. Thus it reports extremely slow speed for prompt rate. It's the same with API. ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.4.0
{ "login": "jessegross", "id": 6468499, "node_id": "MDQ6VXNlcjY0Njg0OTk=", "avatar_url": "https://avatars.githubusercontent.com/u/6468499?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jessegross", "html_url": "https://github.com/jessegross", "followers_url": "https://api.github.com/users/jessegross/followers", "following_url": "https://api.github.com/users/jessegross/following{/other_user}", "gists_url": "https://api.github.com/users/jessegross/gists{/gist_id}", "starred_url": "https://api.github.com/users/jessegross/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jessegross/subscriptions", "organizations_url": "https://api.github.com/users/jessegross/orgs", "repos_url": "https://api.github.com/users/jessegross/repos", "events_url": "https://api.github.com/users/jessegross/events{/privacy}", "received_events_url": "https://api.github.com/users/jessegross/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7592/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7592/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4335
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4335/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4335/comments
https://api.github.com/repos/ollama/ollama/issues/4335/events
https://github.com/ollama/ollama/issues/4335
2,290,622,773
I_kwDOJ0Z1Ps6IiCE1
4,335
AMD Gpu RX 6600 on WIndows 10
{ "login": "usmandilmeer", "id": 51738693, "node_id": "MDQ6VXNlcjUxNzM4Njkz", "avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4", "gravatar_id": "", "url": "https://api.github.com/users/usmandilmeer", "html_url": "https://github.com/usmandilmeer", "followers_url": "https://api.github.com/users/usmandilmeer/followers", "following_url": "https://api.github.com/users/usmandilmeer/following{/other_user}", "gists_url": "https://api.github.com/users/usmandilmeer/gists{/gist_id}", "starred_url": "https://api.github.com/users/usmandilmeer/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/usmandilmeer/subscriptions", "organizations_url": "https://api.github.com/users/usmandilmeer/orgs", "repos_url": "https://api.github.com/users/usmandilmeer/repos", "events_url": "https://api.github.com/users/usmandilmeer/events{/privacy}", "received_events_url": "https://api.github.com/users/usmandilmeer/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-05-11T03:06:41
2024-05-11T04:47:20
2024-05-11T04:23:49
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm having trouble using my AMD GPU RX 6600 on Windows 10 with Ollama. Could someone please assist me and provide any workarounds to make this AMD GPU RX 6600 function? Please provide step-by-step instructions specifically for Windows 10. Thank you. ### OS Windows ### GPU AMD ### CPU Intel ### Ollama version 0.1.35
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4335/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4335/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/60
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/60/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/60/comments
https://api.github.com/repos/ollama/ollama/issues/60/events
https://github.com/ollama/ollama/issues/60
1,795,167,094
I_kwDOJ0Z1Ps5rABN2
60
crash on large context sizes
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2023-07-08T23:43:30
2023-07-27T23:46:30
2023-07-27T23:46:30
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
For prompts larger than ~100 tokens the server will sometimes crash
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/60/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/60/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2012
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2012/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2012/comments
https://api.github.com/repos/ollama/ollama/issues/2012/events
https://github.com/ollama/ollama/issues/2012
2,083,143,527
I_kwDOJ0Z1Ps58Kj9n
2,012
Ollama not using my gpu whatsoever.
{ "login": "Motzumoto", "id": 45925152, "node_id": "MDQ6VXNlcjQ1OTI1MTUy", "avatar_url": "https://avatars.githubusercontent.com/u/45925152?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Motzumoto", "html_url": "https://github.com/Motzumoto", "followers_url": "https://api.github.com/users/Motzumoto/followers", "following_url": "https://api.github.com/users/Motzumoto/following{/other_user}", "gists_url": "https://api.github.com/users/Motzumoto/gists{/gist_id}", "starred_url": "https://api.github.com/users/Motzumoto/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Motzumoto/subscriptions", "organizations_url": "https://api.github.com/users/Motzumoto/orgs", "repos_url": "https://api.github.com/users/Motzumoto/repos", "events_url": "https://api.github.com/users/Motzumoto/events{/privacy}", "received_events_url": "https://api.github.com/users/Motzumoto/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
26
2024-01-16T07:03:29
2024-03-11T19:37:28
2024-03-11T19:37:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
![image](https://github.com/jmorganca/ollama/assets/45925152/368ba9e2-8113-46e7-9192-43f27ff91fb9) I do have cuda drivers installed: ![image](https://github.com/jmorganca/ollama/assets/45925152/bbd87158-7f01-40ee-98b9-c111858cd238)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2012/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2012/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/407
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/407/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/407/comments
https://api.github.com/repos/ollama/ollama/issues/407/events
https://github.com/ollama/ollama/pull/407
1,865,961,586
PR_kwDOJ0Z1Ps5YvLqL
407
Build Automation
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-08-24T22:13:18
2023-09-14T20:33:47
2023-09-12T19:26:53
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/407", "html_url": "https://github.com/ollama/ollama/pull/407", "diff_url": "https://github.com/ollama/ollama/pull/407.diff", "patch_url": "https://github.com/ollama/ollama/pull/407.patch", "merged_at": null }
Working on moving our builds to a github runner on release. This is a CPU amd64 build. No arm64 in this PR because github doesn't have an arm64 runner, but we need to build the llama.cpp exe on an arm64 machine for the arm64 release. We can use an external build runner to accomplish this.
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/407/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 0, "hooray": 3, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/407/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8249
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8249/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8249/comments
https://api.github.com/repos/ollama/ollama/issues/8249/events
https://github.com/ollama/ollama/issues/8249
2,759,876,827
I_kwDOJ0Z1Ps6kgGDb
8,249
DeepSeek v3
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/followers", "following_url": "https://api.github.com/users/olumolu/following{/other_user}", "gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}", "starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/olumolu/subscriptions", "organizations_url": "https://api.github.com/users/olumolu/orgs", "repos_url": "https://api.github.com/users/olumolu/repos", "events_url": "https://api.github.com/users/olumolu/events{/privacy}", "received_events_url": "https://api.github.com/users/olumolu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
45
2024-12-26T15:13:45
2025-01-15T15:30:59
2025-01-14T00:26:00
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
https://huggingface.co/deepseek-ai/DeepSeek-V3-Base https://huggingface.co/deepseek-ai/DeepSeek-V3 Kindly make it available to ollama so we can test.. Model paper https://github.com/deepseek-ai/DeepSeek-V3/blob/main/DeepSeek_V3.pdf https://x.com/deepseek_ai/status/1872242657348710721
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8249/reactions", "total_count": 72, "+1": 69, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 3 }
https://api.github.com/repos/ollama/ollama/issues/8249/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1957
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1957/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1957/comments
https://api.github.com/repos/ollama/ollama/issues/1957/events
https://github.com/ollama/ollama/pull/1957
2,079,349,444
PR_kwDOJ0Z1Ps5j89Is
1,957
Add MindMac to Community Integrations -> Web & Desktop section
{ "login": "hugo53", "id": 1033299, "node_id": "MDQ6VXNlcjEwMzMyOTk=", "avatar_url": "https://avatars.githubusercontent.com/u/1033299?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hugo53", "html_url": "https://github.com/hugo53", "followers_url": "https://api.github.com/users/hugo53/followers", "following_url": "https://api.github.com/users/hugo53/following{/other_user}", "gists_url": "https://api.github.com/users/hugo53/gists{/gist_id}", "starred_url": "https://api.github.com/users/hugo53/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hugo53/subscriptions", "organizations_url": "https://api.github.com/users/hugo53/orgs", "repos_url": "https://api.github.com/users/hugo53/repos", "events_url": "https://api.github.com/users/hugo53/events{/privacy}", "received_events_url": "https://api.github.com/users/hugo53/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-01-12T17:31:14
2024-01-31T15:53:07
2024-01-31T15:48:37
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1957", "html_url": "https://github.com/ollama/ollama/pull/1957", "diff_url": "https://github.com/ollama/ollama/pull/1957.diff", "patch_url": "https://github.com/ollama/ollama/pull/1957.patch", "merged_at": "2024-01-31T15:48:37" }
Hi there, MindMac is a privacy-first & feature-rich GPT client for macOS, designed for maximum productivity. It already has Ollama support, enabling users to run any model on their devices and easily connect with MindMac to ask questions seamlessly. Quick documentation can be found [here](https://docs.mindmac.app/how-to.../add-ollama-endpoint). Please help to review this PR. Thank you in advance. Best regards, Hoang
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1957/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1957/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4616
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4616/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4616/comments
https://api.github.com/repos/ollama/ollama/issues/4616/events
https://github.com/ollama/ollama/issues/4616
2,315,898,872
I_kwDOJ0Z1Ps6KCc_4
4,616
Ollama downloading wrong driver.
{ "login": "Murat-U-Saglam", "id": 42717642, "node_id": "MDQ6VXNlcjQyNzE3NjQy", "avatar_url": "https://avatars.githubusercontent.com/u/42717642?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Murat-U-Saglam", "html_url": "https://github.com/Murat-U-Saglam", "followers_url": "https://api.github.com/users/Murat-U-Saglam/followers", "following_url": "https://api.github.com/users/Murat-U-Saglam/following{/other_user}", "gists_url": "https://api.github.com/users/Murat-U-Saglam/gists{/gist_id}", "starred_url": "https://api.github.com/users/Murat-U-Saglam/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Murat-U-Saglam/subscriptions", "organizations_url": "https://api.github.com/users/Murat-U-Saglam/orgs", "repos_url": "https://api.github.com/users/Murat-U-Saglam/repos", "events_url": "https://api.github.com/users/Murat-U-Saglam/events{/privacy}", "received_events_url": "https://api.github.com/users/Murat-U-Saglam/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 6677745918, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g", "url": "https://api.github.com/repos/ollama/ollama/labels/gpu", "name": "gpu", "color": "76C49E", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-05-24T17:32:21
2024-06-18T13:41:19
2024-06-18T13:40:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Ollama is downloading NVIDIA drivers even though I have AMD GPU. I have ROCM installed ``` apt show rocm-libs -a Package: rocm-libs Version: 6.1.1.60101-90~22.04 ``` and I use the following script to setup [Ollama](https://ollama.com/download/linux) ` curl -fsSL https://ollama.com/install.sh | sh` The output of running the script ``` curl -fsSL https://ollama.com/install.sh | sh >>> Downloading ollama... ######################################################################## 100.0%##O=# # ######################################################################## 100.0% >>> Installing ollama to /usr/local/bin... [sudo] password for ----: >>> Creating ollama user... >>> Adding ollama user to render group... >>> Adding ollama user to video group... >>> Adding current user to ollama group... >>> Creating ollama systemd service... >>> Enabling and starting ollama service... Created symlink /etc/systemd/system/default.target.wants/ollama.service → /etc/systemd/system/ollama.service. >>> NVIDIA GPU installed. ``` ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version N/A
{ "login": "Murat-U-Saglam", "id": 42717642, "node_id": "MDQ6VXNlcjQyNzE3NjQy", "avatar_url": "https://avatars.githubusercontent.com/u/42717642?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Murat-U-Saglam", "html_url": "https://github.com/Murat-U-Saglam", "followers_url": "https://api.github.com/users/Murat-U-Saglam/followers", "following_url": "https://api.github.com/users/Murat-U-Saglam/following{/other_user}", "gists_url": "https://api.github.com/users/Murat-U-Saglam/gists{/gist_id}", "starred_url": "https://api.github.com/users/Murat-U-Saglam/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Murat-U-Saglam/subscriptions", "organizations_url": "https://api.github.com/users/Murat-U-Saglam/orgs", "repos_url": "https://api.github.com/users/Murat-U-Saglam/repos", "events_url": "https://api.github.com/users/Murat-U-Saglam/events{/privacy}", "received_events_url": "https://api.github.com/users/Murat-U-Saglam/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4616/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4616/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/218
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/218/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/218/comments
https://api.github.com/repos/ollama/ollama/issues/218/events
https://github.com/ollama/ollama/issues/218
1,822,635,478
I_kwDOJ0Z1Ps5sozXW
218
CLI doesn't install properly if app isn't named `Ollama.app`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5675428184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUkgpWA", "url": "https://api.github.com/repos/ollama/ollama/labels/app", "name": "app", "color": "000000", "default": false, "description": "" } ]
closed
false
null
[]
null
0
2023-07-26T15:14:05
2023-07-27T14:53:42
2023-07-27T14:53:42
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
If the app filename has a space in it (e.g. `Ollama 2.app`), the cli won't install itself in `/usr/local/bin/ollama` properly
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/218/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/218/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2827
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2827/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2827/comments
https://api.github.com/repos/ollama/ollama/issues/2827/events
https://github.com/ollama/ollama/pull/2827
2,160,454,556
PR_kwDOJ0Z1Ps5oQGF_
2,827
fix: print usedMemory size right
{ "login": "tylinux", "id": 1974874, "node_id": "MDQ6VXNlcjE5NzQ4NzQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1974874?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tylinux", "html_url": "https://github.com/tylinux", "followers_url": "https://api.github.com/users/tylinux/followers", "following_url": "https://api.github.com/users/tylinux/following{/other_user}", "gists_url": "https://api.github.com/users/tylinux/gists{/gist_id}", "starred_url": "https://api.github.com/users/tylinux/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tylinux/subscriptions", "organizations_url": "https://api.github.com/users/tylinux/orgs", "repos_url": "https://api.github.com/users/tylinux/repos", "events_url": "https://api.github.com/users/tylinux/events{/privacy}", "received_events_url": "https://api.github.com/users/tylinux/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-29T06:02:07
2024-02-29T19:11:12
2024-02-29T19:11:05
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2827", "html_url": "https://github.com/ollama/ollama/pull/2827", "diff_url": "https://github.com/ollama/ollama/pull/2827.diff", "patch_url": "https://github.com/ollama/ollama/pull/2827.patch", "merged_at": "2024-02-29T19:11:04" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2827/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2827/timeline
null
null
true