url
stringlengths
51
54
repository_url
stringclasses
1 value
labels_url
stringlengths
65
68
comments_url
stringlengths
60
63
events_url
stringlengths
58
61
html_url
stringlengths
39
44
id
int64
1.78B
2.82B
node_id
stringlengths
18
19
number
int64
1
8.69k
title
stringlengths
1
382
user
dict
labels
listlengths
0
5
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
2
milestone
null
comments
int64
0
323
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
2
118k
closed_by
dict
reactions
dict
timeline_url
stringlengths
60
63
performed_via_github_app
null
state_reason
stringclasses
4 values
is_pull_request
bool
2 classes
https://api.github.com/repos/ollama/ollama/issues/8250
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8250/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8250/comments
https://api.github.com/repos/ollama/ollama/issues/8250/events
https://github.com/ollama/ollama/issues/8250
2,759,878,106
I_kwDOJ0Z1Ps6kgGXa
8,250
qwen qvq model
{ "login": "olumolu", "id": 162728301, "node_id": "U_kgDOCbMJbQ", "avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4", "gravatar_id": "", "url": "https://api.github.com/users/olumolu", "html_url": "https://github.com/olumolu", "followers_url": "https://api.github.com/users/olumolu/followers", "following_url": "https://api.github.com/users/olumolu/following{/other_user}", "gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}", "starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/olumolu/subscriptions", "organizations_url": "https://api.github.com/users/olumolu/orgs", "repos_url": "https://api.github.com/users/olumolu/repos", "events_url": "https://api.github.com/users/olumolu/events{/privacy}", "received_events_url": "https://api.github.com/users/olumolu/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-12-26T15:15:15
2025-01-16T08:57:20
2024-12-29T19:16:56
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Extremely good with reasoning and maths mainly better than openai o1 https://huggingface.co/Qwen/QVQ-72B-Preview
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8250/reactions", "total_count": 9, "+1": 9, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8250/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5493
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5493/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5493/comments
https://api.github.com/repos/ollama/ollama/issues/5493/events
https://github.com/ollama/ollama/issues/5493
2,391,664,217
I_kwDOJ0Z1Ps6OjeZZ
5,493
unable to load nvcuda
{ "login": "yake-cyber", "id": 174697336, "node_id": "U_kgDOCmmreA", "avatar_url": "https://avatars.githubusercontent.com/u/174697336?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yake-cyber", "html_url": "https://github.com/yake-cyber", "followers_url": "https://api.github.com/users/yake-cyber/followers", "following_url": "https://api.github.com/users/yake-cyber/following{/other_user}", "gists_url": "https://api.github.com/users/yake-cyber/gists{/gist_id}", "starred_url": "https://api.github.com/users/yake-cyber/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yake-cyber/subscriptions", "organizations_url": "https://api.github.com/users/yake-cyber/orgs", "repos_url": "https://api.github.com/users/yake-cyber/repos", "events_url": "https://api.github.com/users/yake-cyber/events{/privacy}", "received_events_url": "https://api.github.com/users/yake-cyber/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
7
2024-07-05T02:38:20
2024-08-13T03:39:30
2024-07-11T08:34:16
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? my ollama does not run on the NVIDIA gpu and i use the debug mode and find this message "time=2024-07-04T17:13:20.134+08:00 level=DEBUG source=gpu.go:385 msg="Unable to load nvcuda" library=/usr/lib/libcuda.so.418.74 error="Unable to load /usr/lib/libcuda.so.418.74 library to query for Nvidia GPUs: /usr/lib/libcuda.so.418.74: wrong ELF class: ELFCLASS32" dlerr: /usr/lib64/libcuda.so.418.74: undefined symbol: cuCtxCreate_v3 time=2024-07-04T17:13:20.135+08:00 level=DEBUG source=gpu.go:385 msg="Unable to load nvcuda" library=/usr/lib64/libcuda.so.418.74 error="symbol lookup for cuCtxCreate_v3 failed: /usr/lib64/libcuda.so.418.74: undefined symbol: cuCtxCreate_v3" time=2024-07-04T17:13:20.135+08:00 level=DEBUG source=gpu.go:304 msg="Searching for GPU library" name=libcudart.so*" could you please clarify what this means and how can i resolving this issue? [debug-ollama.txt](https://github.com/user-attachments/files/16104245/debug-ollama.txt) ![nvidia-smi](https://github.com/ollama/ollama/assets/174697336/16f66b47-fa5c-4be6-a748-7e5173e69279) [log20240704.txt](https://github.com/user-attachments/files/16104248/log20240704.txt) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.40
{ "login": "yake-cyber", "id": 174697336, "node_id": "U_kgDOCmmreA", "avatar_url": "https://avatars.githubusercontent.com/u/174697336?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yake-cyber", "html_url": "https://github.com/yake-cyber", "followers_url": "https://api.github.com/users/yake-cyber/followers", "following_url": "https://api.github.com/users/yake-cyber/following{/other_user}", "gists_url": "https://api.github.com/users/yake-cyber/gists{/gist_id}", "starred_url": "https://api.github.com/users/yake-cyber/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yake-cyber/subscriptions", "organizations_url": "https://api.github.com/users/yake-cyber/orgs", "repos_url": "https://api.github.com/users/yake-cyber/repos", "events_url": "https://api.github.com/users/yake-cyber/events{/privacy}", "received_events_url": "https://api.github.com/users/yake-cyber/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5493/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5493/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6252
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6252/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6252/comments
https://api.github.com/repos/ollama/ollama/issues/6252/events
https://github.com/ollama/ollama/issues/6252
2,454,796,873
I_kwDOJ0Z1Ps6SUTpJ
6,252
cross compiling issue
{ "login": "andyyumiao", "id": 11346379, "node_id": "MDQ6VXNlcjExMzQ2Mzc5", "avatar_url": "https://avatars.githubusercontent.com/u/11346379?v=4", "gravatar_id": "", "url": "https://api.github.com/users/andyyumiao", "html_url": "https://github.com/andyyumiao", "followers_url": "https://api.github.com/users/andyyumiao/followers", "following_url": "https://api.github.com/users/andyyumiao/following{/other_user}", "gists_url": "https://api.github.com/users/andyyumiao/gists{/gist_id}", "starred_url": "https://api.github.com/users/andyyumiao/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/andyyumiao/subscriptions", "organizations_url": "https://api.github.com/users/andyyumiao/orgs", "repos_url": "https://api.github.com/users/andyyumiao/repos", "events_url": "https://api.github.com/users/andyyumiao/events{/privacy}", "received_events_url": "https://api.github.com/users/andyyumiao/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-08-08T03:31:23
2024-08-10T00:04:39
2024-08-10T00:04:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? **When cross platform cross compiling, for example, when I compile a Linux version of a program on a Mac, the following error is reported:** `gpu/amd_linux.go:200:19: undefined: RocmComputeMin gpu/amd_linux.go:273:20: undefined: IGPUMemLimit gpu/amd_linux.go:295:20: undefined: rocmMinimumMemory` <img width="739" alt="1111" src="https://github.com/user-attachments/assets/7525b048-97fc-4e75-8731-695f8643ee6b"> **My execution steps are:** `export GOOS=linux` `export GOARCH=amd64` `go generate ./...` `go build` **### OS** macOS **### GPU** Apple **### CPU** Apple **### Ollama version** the latest master branch
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6252/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6252/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6014
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6014/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6014/comments
https://api.github.com/repos/ollama/ollama/issues/6014/events
https://github.com/ollama/ollama/pull/6014
2,433,401,119
PR_kwDOJ0Z1Ps52pC1G
6,014
server: add OLLAMA_RUNNERS_DIR to help description
{ "login": "jing-rui", "id": 51155955, "node_id": "MDQ6VXNlcjUxMTU1OTU1", "avatar_url": "https://avatars.githubusercontent.com/u/51155955?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jing-rui", "html_url": "https://github.com/jing-rui", "followers_url": "https://api.github.com/users/jing-rui/followers", "following_url": "https://api.github.com/users/jing-rui/following{/other_user}", "gists_url": "https://api.github.com/users/jing-rui/gists{/gist_id}", "starred_url": "https://api.github.com/users/jing-rui/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jing-rui/subscriptions", "organizations_url": "https://api.github.com/users/jing-rui/orgs", "repos_url": "https://api.github.com/users/jing-rui/repos", "events_url": "https://api.github.com/users/jing-rui/events{/privacy}", "received_events_url": "https://api.github.com/users/jing-rui/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-27T09:55:05
2024-11-23T19:31:15
2024-11-23T19:31:15
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6014", "html_url": "https://github.com/ollama/ollama/pull/6014", "diff_url": "https://github.com/ollama/ollama/pull/6014.diff", "patch_url": "https://github.com/ollama/ollama/pull/6014.patch", "merged_at": null }
The env OLLAMA_RUNNERS_DIR is useful to avoid extract embedded files every time at `ollama serve` start.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6014/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6014/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7474
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7474/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7474/comments
https://api.github.com/repos/ollama/ollama/issues/7474/events
https://github.com/ollama/ollama/pull/7474
2,630,850,182
PR_kwDOJ0Z1Ps6AtjcF
7,474
Fix: return direct URL when OCI registry is not redirecting
{ "login": "peterwilli", "id": 1212814, "node_id": "MDQ6VXNlcjEyMTI4MTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1212814?v=4", "gravatar_id": "", "url": "https://api.github.com/users/peterwilli", "html_url": "https://github.com/peterwilli", "followers_url": "https://api.github.com/users/peterwilli/followers", "following_url": "https://api.github.com/users/peterwilli/following{/other_user}", "gists_url": "https://api.github.com/users/peterwilli/gists{/gist_id}", "starred_url": "https://api.github.com/users/peterwilli/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/peterwilli/subscriptions", "organizations_url": "https://api.github.com/users/peterwilli/orgs", "repos_url": "https://api.github.com/users/peterwilli/repos", "events_url": "https://api.github.com/users/peterwilli/events{/privacy}", "received_events_url": "https://api.github.com/users/peterwilli/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2024-11-02T23:00:57
2025-01-09T07:46:52
2024-11-21T11:04:10
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7474", "html_url": "https://github.com/ollama/ollama/pull/7474", "diff_url": "https://github.com/ollama/ollama/pull/7474.diff", "patch_url": "https://github.com/ollama/ollama/pull/7474.patch", "merged_at": null }
When debugging why I couldn't use a regular OCI registry (See screenshot below), I found out the ollama server was doing a redirect to AWS. This change returns the direct URL when such redirect does not happen, while assuring regular behavior when it does, so that regular pulls still work! <img width="1710" alt="SCR-20241102-rgnu" src="https://github.com/user-attachments/assets/26c97b1e-f0d6-4b9f-82ac-a33e14f64fb3"> This PR fixes https://github.com/ollama/ollama/issues/7244
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7474/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7474/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3712
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3712/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3712/comments
https://api.github.com/repos/ollama/ollama/issues/3712/events
https://github.com/ollama/ollama/pull/3712
2,249,256,817
PR_kwDOJ0Z1Ps5s-ZFw
3,712
add stablelm graph calculation
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-04-17T20:57:31
2024-04-17T22:57:51
2024-04-17T22:57:51
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3712", "html_url": "https://github.com/ollama/ollama/pull/3712", "diff_url": "https://github.com/ollama/ollama/pull/3712.diff", "patch_url": "https://github.com/ollama/ollama/pull/3712.patch", "merged_at": "2024-04-17T22:57:51" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3712/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3712/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3102
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3102/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3102/comments
https://api.github.com/repos/ollama/ollama/issues/3102/events
https://github.com/ollama/ollama/issues/3102
2,183,879,753
I_kwDOJ0Z1Ps6CK1xJ
3,102
Response_format not supported
{ "login": "halcwb", "id": 683631, "node_id": "MDQ6VXNlcjY4MzYzMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/683631?v=4", "gravatar_id": "", "url": "https://api.github.com/users/halcwb", "html_url": "https://github.com/halcwb", "followers_url": "https://api.github.com/users/halcwb/followers", "following_url": "https://api.github.com/users/halcwb/following{/other_user}", "gists_url": "https://api.github.com/users/halcwb/gists{/gist_id}", "starred_url": "https://api.github.com/users/halcwb/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/halcwb/subscriptions", "organizations_url": "https://api.github.com/users/halcwb/orgs", "repos_url": "https://api.github.com/users/halcwb/repos", "events_url": "https://api.github.com/users/halcwb/events{/privacy}", "received_events_url": "https://api.github.com/users/halcwb/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
2
2024-03-13T12:25:32
2024-03-13T15:01:10
2024-03-13T15:01:10
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
When sending this request to the open AI endpoint I don't get the requested JSON. The payload is the actual content send to the api. Exactly the same content works with for example the fireworks api. ℹ INFO: EndPoint: http://localhost:11434/api/chat Payload: {"format":"json","messages":[{"content":"What is the minimal age for a neonate 28 weeks to 32 weeks corrected gestational age. Reply just in one JSON.","role":"user"}],"model":"llama2","options":{"num_keep":null,"seed":101,"num_predict":null,"top_k":null,"top_p":null,"tfs_z":null,"typical_p":null,"repeat_last_n":64,"temperature":0.0,"repeat_penalty":null,"presence_penalty":null,"frequency_penalty":null,"mirostat":0,"mirostat_tau":null,"mirostat_eta":null,"penalize_newline":null,"stop":[],"numa":null,"num_ctx":2048,"num_batch":null,"num_gqa":null,"num_gpu":null,"main_gpu":null,"low_vram":null,"f16_kv":null,"vocab_only":null,"use_mmap":null,"use_mlock":null,"rope_frequency_base":null,"rope_frequency_scale":null,"num_thread":null},"response_format":{"schema":{ "$schema": "http://json-schema.org/draft-04/schema#", "title": "<>f__AnonymousType3408661278OfIntegerAndString", "type": "object", "additionalProperties": false, "properties": { "number": { "type": "integer", "format": "int32" }, "unit": { "type": [ "null", "string" ] } } },"type":"json_object"},"stream":false} val it: Result<Response<Ollama.OllamaResponse>,string> = Ok { Original = "{"model":"llama2","created_at":"2024-03-13T12:22:59.29158Z","message":{"role":"assistant","content":"{\n\"minimal age\": 4000,\n\"units\": \"days\"\n}\n\n\n\n"},"done":true,"total_duration":809391791,"load_duration":2388625,"prompt_eval_count":50,"prompt_eval_duration":341193000,"eval_count":28,"eval_duration":464799000}" Response = { error = null model = "llama2" created_at = "2024-03-13T12:22:59.29158Z" response = null message = { content = "{ "minimal age": 4000, "units": "days" } " role = "assistant" } done = true context = null total_duration = 809391791L load_duration = 2388625L prompt_eval_duration = 341193000L eval_count = 28 eval_duration = 464799000L } } The workaround mentioned in #2844 I don't get to work
{ "login": "halcwb", "id": 683631, "node_id": "MDQ6VXNlcjY4MzYzMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/683631?v=4", "gravatar_id": "", "url": "https://api.github.com/users/halcwb", "html_url": "https://github.com/halcwb", "followers_url": "https://api.github.com/users/halcwb/followers", "following_url": "https://api.github.com/users/halcwb/following{/other_user}", "gists_url": "https://api.github.com/users/halcwb/gists{/gist_id}", "starred_url": "https://api.github.com/users/halcwb/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/halcwb/subscriptions", "organizations_url": "https://api.github.com/users/halcwb/orgs", "repos_url": "https://api.github.com/users/halcwb/repos", "events_url": "https://api.github.com/users/halcwb/events{/privacy}", "received_events_url": "https://api.github.com/users/halcwb/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3102/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3102/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1253
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1253/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1253/comments
https://api.github.com/repos/ollama/ollama/issues/1253/events
https://github.com/ollama/ollama/issues/1253
2,007,582,639
I_kwDOJ0Z1Ps53qUev
1,253
Error when downloading and running any dataset of any size.
{ "login": "ll3N1GmAll", "id": 10640635, "node_id": "MDQ6VXNlcjEwNjQwNjM1", "avatar_url": "https://avatars.githubusercontent.com/u/10640635?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ll3N1GmAll", "html_url": "https://github.com/ll3N1GmAll", "followers_url": "https://api.github.com/users/ll3N1GmAll/followers", "following_url": "https://api.github.com/users/ll3N1GmAll/following{/other_user}", "gists_url": "https://api.github.com/users/ll3N1GmAll/gists{/gist_id}", "starred_url": "https://api.github.com/users/ll3N1GmAll/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ll3N1GmAll/subscriptions", "organizations_url": "https://api.github.com/users/ll3N1GmAll/orgs", "repos_url": "https://api.github.com/users/ll3N1GmAll/repos", "events_url": "https://api.github.com/users/ll3N1GmAll/events{/privacy}", "received_events_url": "https://api.github.com/users/ll3N1GmAll/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
4
2023-11-23T06:51:02
2023-11-24T01:11:28
2023-11-23T17:27:07
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
This is the error I get after d/l a dataset and when trying to run a dataset - "Error: llama runner process has terminated" It pulls them down, verifies the hash, then says "success", the very next line is the error above. I am running Xubuntu 22.04, 16GB RAM, Intel Pentium CPU G4560 @ 3.50GHz, 8x Nvidia 1080Ti GPUs. I get this with even small sets like the 1.8GB starcoder set. After a reboot, trying to run a dataset with "ollama run <dataset-name>" results in several seconds of attempting to start. The "ollama serve" process is visible in task manager, then the error "Error: llama runner process has terminated" is displayed in the terminal. The "ollama serve" process remains running/hung in task manager, consuming roughly 400MB of RAM. Which is the amount it was consuming while the terminal process was trying to run the dataset. Manually killing the process and trying to run it again results in the exact same behavior as after a reboot, with the exception that is fails to the error within a second or so tops instead of taking several seconds to fail like it did after the reboot. It still is consuming ~400MB of RAM. This looks similar to issue #788; but the newer version is supposed to prevent the AVX CPU requirement from causing this issue. However, I still have this issue. results of "journalctl -u ollama" Nov 22 23:07:12 <machine-name> systemd[1]: Started Ollama Service. Nov 22 23:07:14 <machine-name> ollama[1572]: 2023/11/22 23:07:14 images.go:779: total blobs: 0 Nov 22 23:07:14 <machine-name> ollama[1572]: 2023/11/22 23:07:14 images.go:786: total unused blobs removed: 0 Nov 22 23:07:14 <machine-name> ollama[1572]: 2023/11/22 23:07:14 routes.go:777: Listening on 127.0.0.1:11434 (version 0.1.11) Nov 22 23:15:09 <machine-name> systemd[1]: Stopping Ollama Service... Nov 22 23:15:09 <machine-name> systemd[1]: ollama.service: Deactivated successfully. Nov 22 23:15:09 <machine-name> systemd[1]: Stopped Ollama Service. Nov 22 23:15:09 <machine-name> systemd[1]: Started Ollama Service. Nov 22 23:15:09 <machine-name> ollama[30889]: 2023/11/22 23:15:09 images.go:779: total blobs: 0 Nov 22 23:15:09 <machine-name> ollama[30889]: 2023/11/22 23:15:09 images.go:786: total unused blobs removed: 0 Nov 22 23:15:09 <machine-name> ollama[30889]: 2023/11/22 23:15:09 routes.go:777: Listening on 127.0.0.1:11434 (version 0.1.11) Nov 22 23:16:17 <machine-name> ollama[30889]: [GIN] 2023/11/22 - 23:16:17 | 200 | 93.552µs | 127.0.0.1 | HEAD "/" Nov 22 23:16:17 <machine-name> ollama[30889]: [GIN] 2023/11/22 - 23:16:17 | 404 | 173.127µs | 127.0.0.1 | POST "/api/show" Nov 22 23:16:19 <machine-name> ollama[30889]: 2023/11/22 23:16:19 download.go:123: downloading 6ae280299950 in 42 100 MB part(s) Nov 22 23:17:18 <machine-name> ollama[30889]: 2023/11/22 23:17:18 download.go:123: downloading 22e1b2e8dc2f in 1 43 B part(s) Nov 22 23:17:21 <machine-name> ollama[30889]: 2023/11/22 23:17:21 download.go:123: downloading e35ab70a78c7 in 1 90 B part(s) Nov 22 23:17:24 <machine-name> ollama[30889]: 2023/11/22 23:17:24 download.go:123: downloading 1cb90d66f4d4 in 1 381 B part(s) Nov 22 23:17:47 <machine-name> ollama[30889]: [GIN] 2023/11/22 - 23:17:47 | 200 | 1m30s | 127.0.0.1 | POST "/api/pull" Nov 22 23:17:47 <machine-name> ollama[30889]: 2023/11/22 23:17:47 llama.go:291: 89320 MB VRAM available, loading up to 546 GPU layers Nov 22 23:17:47 <machine-name> ollama[30889]: 2023/11/22 23:17:47 llama.go:420: starting llama runner Nov 22 23:17:47 <machine-name> ollama[30889]: 2023/11/22 23:17:47 llama.go:478: waiting for llama runner to start responding Nov 22 23:17:48 <machine-name> ollama[30889]: 2023/11/22 23:17:48 llama.go:435: signal: illegal instruction (core dumped) Nov 22 23:17:48 <machine-name> ollama[30889]: 2023/11/22 23:17:48 llama.go:443: error starting llama runner: llama runner process has terminated Nov 22 23:17:48 <machine-name> ollama[30889]: 2023/11/22 23:17:48 llama.go:509: llama runner stopped successfully Nov 22 23:17:48 <machine-name> ollama[30889]: 2023/11/22 23:17:48 llama.go:420: starting llama runner Nov 22 23:17:48 <machine-name> ollama[30889]: 2023/11/22 23:17:48 llama.go:478: waiting for llama runner to start responding
{ "login": "ll3N1GmAll", "id": 10640635, "node_id": "MDQ6VXNlcjEwNjQwNjM1", "avatar_url": "https://avatars.githubusercontent.com/u/10640635?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ll3N1GmAll", "html_url": "https://github.com/ll3N1GmAll", "followers_url": "https://api.github.com/users/ll3N1GmAll/followers", "following_url": "https://api.github.com/users/ll3N1GmAll/following{/other_user}", "gists_url": "https://api.github.com/users/ll3N1GmAll/gists{/gist_id}", "starred_url": "https://api.github.com/users/ll3N1GmAll/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ll3N1GmAll/subscriptions", "organizations_url": "https://api.github.com/users/ll3N1GmAll/orgs", "repos_url": "https://api.github.com/users/ll3N1GmAll/repos", "events_url": "https://api.github.com/users/ll3N1GmAll/events{/privacy}", "received_events_url": "https://api.github.com/users/ll3N1GmAll/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1253/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1253/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8688
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8688/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8688/comments
https://api.github.com/repos/ollama/ollama/issues/8688/events
https://github.com/ollama/ollama/pull/8688
2,820,160,395
PR_kwDOJ0Z1Ps6JduvO
8,688
Add library in Zig.
{ "login": "dravenk", "id": 14295318, "node_id": "MDQ6VXNlcjE0Mjk1MzE4", "avatar_url": "https://avatars.githubusercontent.com/u/14295318?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dravenk", "html_url": "https://github.com/dravenk", "followers_url": "https://api.github.com/users/dravenk/followers", "following_url": "https://api.github.com/users/dravenk/following{/other_user}", "gists_url": "https://api.github.com/users/dravenk/gists{/gist_id}", "starred_url": "https://api.github.com/users/dravenk/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dravenk/subscriptions", "organizations_url": "https://api.github.com/users/dravenk/orgs", "repos_url": "https://api.github.com/users/dravenk/repos", "events_url": "https://api.github.com/users/dravenk/events{/privacy}", "received_events_url": "https://api.github.com/users/dravenk/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2025-01-30T08:05:43
2025-01-30T08:05:43
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8688", "html_url": "https://github.com/ollama/ollama/pull/8688", "diff_url": "https://github.com/ollama/ollama/pull/8688.diff", "patch_url": "https://github.com/ollama/ollama/pull/8688.patch", "merged_at": null }
null
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8688/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8688/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6440
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6440/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6440/comments
https://api.github.com/repos/ollama/ollama/issues/6440/events
https://github.com/ollama/ollama/issues/6440
2,475,276,834
I_kwDOJ0Z1Ps6Tiboi
6,440
Model architecture Gemma2ForCausalLm
{ "login": "luisgg98", "id": 45603226, "node_id": "MDQ6VXNlcjQ1NjAzMjI2", "avatar_url": "https://avatars.githubusercontent.com/u/45603226?v=4", "gravatar_id": "", "url": "https://api.github.com/users/luisgg98", "html_url": "https://github.com/luisgg98", "followers_url": "https://api.github.com/users/luisgg98/followers", "following_url": "https://api.github.com/users/luisgg98/following{/other_user}", "gists_url": "https://api.github.com/users/luisgg98/gists{/gist_id}", "starred_url": "https://api.github.com/users/luisgg98/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/luisgg98/subscriptions", "organizations_url": "https://api.github.com/users/luisgg98/orgs", "repos_url": "https://api.github.com/users/luisgg98/repos", "events_url": "https://api.github.com/users/luisgg98/events{/privacy}", "received_events_url": "https://api.github.com/users/luisgg98/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-08-20T10:22:23
2024-08-21T20:15:34
2024-08-21T20:15:34
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Good afternoon, I would like to start marking I am not 100% sure whether this is an issue or maybe I am misunderstanding the concept of architecture. I tried to create a model on ollama by using a Modelfile at version 0.3.0. ![imagen](https://github.com/user-attachments/assets/dfdd4b08-5eb8-4939-a536-fd739dd2e784) I go the previous error, so I decided to upgrade ollama to version 0.3.6 Which pops off error "unsupported architecture": ![imagen](https://github.com/user-attachments/assets/b51c374e-dec6-4e81-acf5-b292ce3af43d) Configuration file of the model I have finetuned: ```json { "_name_or_path": "google/gemma-2-9b", "architectures": [ "Gemma2ForCausalLM" ], "attention_bias": false, "attention_dropout": 0.0, "attn_logit_softcapping": 50.0, "bos_token_id": 2, "cache_implementation": "hybrid", "eos_token_id": 1, "final_logit_softcapping": 30.0, "head_dim": 256, "hidden_act": "gelu_pytorch_tanh", "hidden_activation": "gelu_pytorch_tanh", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "model_type": "gemma2", "num_attention_heads": 16, "num_hidden_layers": 42, "num_key_value_heads": 8, "pad_token_id": 0, "query_pre_attn_scalar": 256, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "sliding_window_size": 4096, "torch_dtype": "float16", "transformers_version": "4.43.3", "use_cache": false, "vocab_size": 256000 } ``` It's a model based on Gemma2 9b. Gemma2 works smoothly. ![imagen](https://github.com/user-attachments/assets/965205ba-03df-4273-bb3e-2ad9ce34b807) Modelfile: ```Dockerfile FROM /path/model/ TEMPLATE """<start_of_turn>user/ {{ if .System }}{{ .System }} {{ end }}{{ .Prompt }}<end_of_turn> <start_of_turn>model {{ .Response }}<end_of_turn>""" PARAMETER stop "<start_of_turn>" PARAMETER stop "<end_of_turn>" SYSTEM """Below is an instruction that describes a task. Write a response that appropriately completes the request. Generate a concise summary in Spanish the following input. Here is the input""" ``` I don't understand by Gemma2 is supported but models based on it are not. Thank you for reaching this point. Please if you have any clue about what's the issue please write it down bellow. ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version 0.3.6
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6440/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6440/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8125
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8125/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8125/comments
https://api.github.com/repos/ollama/ollama/issues/8125/events
https://github.com/ollama/ollama/pull/8125
2,743,674,309
PR_kwDOJ0Z1Ps6FbSk2
8,125
darwin: restore multiple runners for x86
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-12-17T00:30:34
2024-12-17T02:45:02
2024-12-17T02:45:02
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8125", "html_url": "https://github.com/ollama/ollama/pull/8125", "diff_url": "https://github.com/ollama/ollama/pull/8125.diff", "patch_url": "https://github.com/ollama/ollama/pull/8125.patch", "merged_at": "2024-12-17T02:45:02" }
In 0.5.2 we simplified packaging to have avx only for macos x86. It looks like there may still be some non-AVX systems out there, so this puts back the prior logic of building no-AVX for the primary binary, and now 2 runners for avx and avx2. These will be packaged in the App bundle only, so the stand-alone binary will now be without AVX support on macos. On arm, we'll also see these runners reported as available in the log, but they're dormant and will never be used at runtime.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8125/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8125/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1214
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1214/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1214/comments
https://api.github.com/repos/ollama/ollama/issues/1214/events
https://github.com/ollama/ollama/issues/1214
2,003,109,082
I_kwDOJ0Z1Ps53ZQTa
1,214
Cache models for system restarts to not download again in docker
{ "login": "peteh", "id": 918728, "node_id": "MDQ6VXNlcjkxODcyOA==", "avatar_url": "https://avatars.githubusercontent.com/u/918728?v=4", "gravatar_id": "", "url": "https://api.github.com/users/peteh", "html_url": "https://github.com/peteh", "followers_url": "https://api.github.com/users/peteh/followers", "following_url": "https://api.github.com/users/peteh/following{/other_user}", "gists_url": "https://api.github.com/users/peteh/gists{/gist_id}", "starred_url": "https://api.github.com/users/peteh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/peteh/subscriptions", "organizations_url": "https://api.github.com/users/peteh/orgs", "repos_url": "https://api.github.com/users/peteh/repos", "events_url": "https://api.github.com/users/peteh/events{/privacy}", "received_events_url": "https://api.github.com/users/peteh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-11-20T22:17:03
2023-11-21T00:07:35
2023-11-21T00:07:35
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I wrote a docker compose file and thought I mapped the right cache folder. However after a system restart, the model is downloaded again. My goal is to map the model cache dir to my local disk so when using the same model after a restart, it is not redownloaded again. The .ollama folder contains a lot of sha256 files which seem to be the downloaded model files but not the final model. My current docker file: ``` version: '3' services: ollama: build: . image: ollama/ollama container_name: ollama volumes: - ./ollama:/root/.ollama ```
{ "login": "peteh", "id": 918728, "node_id": "MDQ6VXNlcjkxODcyOA==", "avatar_url": "https://avatars.githubusercontent.com/u/918728?v=4", "gravatar_id": "", "url": "https://api.github.com/users/peteh", "html_url": "https://github.com/peteh", "followers_url": "https://api.github.com/users/peteh/followers", "following_url": "https://api.github.com/users/peteh/following{/other_user}", "gists_url": "https://api.github.com/users/peteh/gists{/gist_id}", "starred_url": "https://api.github.com/users/peteh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/peteh/subscriptions", "organizations_url": "https://api.github.com/users/peteh/orgs", "repos_url": "https://api.github.com/users/peteh/repos", "events_url": "https://api.github.com/users/peteh/events{/privacy}", "received_events_url": "https://api.github.com/users/peteh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1214/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1214/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4666
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4666/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4666/comments
https://api.github.com/repos/ollama/ollama/issues/4666/events
https://github.com/ollama/ollama/issues/4666
2,319,302,049
I_kwDOJ0Z1Ps6KPb2h
4,666
ollama doesn't create a model from modelfile and gives an error
{ "login": "tMrMorgan", "id": 170948386, "node_id": "U_kgDOCjB3Ig", "avatar_url": "https://avatars.githubusercontent.com/u/170948386?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tMrMorgan", "html_url": "https://github.com/tMrMorgan", "followers_url": "https://api.github.com/users/tMrMorgan/followers", "following_url": "https://api.github.com/users/tMrMorgan/following{/other_user}", "gists_url": "https://api.github.com/users/tMrMorgan/gists{/gist_id}", "starred_url": "https://api.github.com/users/tMrMorgan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tMrMorgan/subscriptions", "organizations_url": "https://api.github.com/users/tMrMorgan/orgs", "repos_url": "https://api.github.com/users/tMrMorgan/repos", "events_url": "https://api.github.com/users/tMrMorgan/events{/privacy}", "received_events_url": "https://api.github.com/users/tMrMorgan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
open
false
null
[]
null
5
2024-05-27T14:32:53
2024-10-26T07:05:55
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Sorry in advance for any mistakes in text when I trying to create a model in terminal, no matter what it based on, and even if the "modelfile" is a stock template of downloaded llm, after command "ollama create test" i got same output everytime " Error: command must be one of "from", "license", "template", "system", "adapter", "parameter", or "message" " i tried to change names of file, change command to e.g. "ollama create test -t modelfile" it's not working at all. also i tried to edit commands in modelfile, so it doesn't get any reaction too. I've read the recent topic with similar error here, as it been suggested - converted modelfile in utf8, download file from topic - didn't helped. And also, i re-installed Ollama. i tried to create modelfile thru Open WebUI from my models, and it works fine. And if i write in dialoge with llm in terminal: "/set system " you are something " and then " /save test " it saves and create new model without any trouble. it is clear that the problem is not with modelfile itself, so what is it? OS is WIN10pro. 16gb RAM, fx8320, gtx1650 with CUDA drivers installed. ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.1.38
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4666/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4666/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8677
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8677/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8677/comments
https://api.github.com/repos/ollama/ollama/issues/8677/events
https://github.com/ollama/ollama/issues/8677
2,819,603,374
I_kwDOJ0Z1Ps6oD7uu
8,677
Wrote scripts to import gguf files/folder
{ "login": "gl2007", "id": 4097227, "node_id": "MDQ6VXNlcjQwOTcyMjc=", "avatar_url": "https://avatars.githubusercontent.com/u/4097227?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gl2007", "html_url": "https://github.com/gl2007", "followers_url": "https://api.github.com/users/gl2007/followers", "following_url": "https://api.github.com/users/gl2007/following{/other_user}", "gists_url": "https://api.github.com/users/gl2007/gists{/gist_id}", "starred_url": "https://api.github.com/users/gl2007/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gl2007/subscriptions", "organizations_url": "https://api.github.com/users/gl2007/orgs", "repos_url": "https://api.github.com/users/gl2007/repos", "events_url": "https://api.github.com/users/gl2007/events{/privacy}", "received_events_url": "https://api.github.com/users/gl2007/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
0
2025-01-30T00:09:02
2025-01-30T00:09:02
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Don't see a "discussion" tab like I see for other repos, so just creating an issue. Had a bunch of gguf's in a folder, so wrote 2 scripts (windows and shell) to import a single gguf and all ggufs in a given folder. Don't know how to get a PR in but I can attach them here is any of you think they are useful.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8677/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8677/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8386
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8386/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8386/comments
https://api.github.com/repos/ollama/ollama/issues/8386/events
https://github.com/ollama/ollama/issues/8386
2,782,119,430
I_kwDOJ0Z1Ps6l08YG
8,386
Return in a response a flag if the input request was truncated
{ "login": "MarkWard0110", "id": 90335263, "node_id": "MDQ6VXNlcjkwMzM1MjYz", "avatar_url": "https://avatars.githubusercontent.com/u/90335263?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MarkWard0110", "html_url": "https://github.com/MarkWard0110", "followers_url": "https://api.github.com/users/MarkWard0110/followers", "following_url": "https://api.github.com/users/MarkWard0110/following{/other_user}", "gists_url": "https://api.github.com/users/MarkWard0110/gists{/gist_id}", "starred_url": "https://api.github.com/users/MarkWard0110/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/MarkWard0110/subscriptions", "organizations_url": "https://api.github.com/users/MarkWard0110/orgs", "repos_url": "https://api.github.com/users/MarkWard0110/repos", "events_url": "https://api.github.com/users/MarkWard0110/events{/privacy}", "received_events_url": "https://api.github.com/users/MarkWard0110/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2025-01-11T19:53:52
2025-01-15T23:59:07
2025-01-15T23:59:07
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Add to the response a flag that indicates true if Ollama truncated the input request. As a developer, I would like the Ollama response to have a flag indicating that it truncated the input prompt so that I can initiate client-side behavior based on this information. In some situations, the accuracy of the chat response depends upon processing the complete input prompt. A client is not informed if the input was truncated. A client may be unable to match Ollama's calculation for the context size required for a given input. The truncation works well for chat conversations, but the user may experience a degradation in the quality of the conversation. The truncation might negatively affect an agent. An agent not interacting with a user might receive a longer-than-expected response in a previous iteration. In the following chat request, Ollama truncates the input prompt. The agent would not be aware the request was truncated. The agent may try to estimate but may not be able to match Ollama's calculation. It also appears that in certain situations, the routes.go chatPrompt call will not truncate the input prompt, but the runner will. I saw this when I had put the context request size one less than the reported prompt eval for the same request. For example, a chat request that the prompt eval is 13. I sent the same request but with a context size of 12. The eval is 55. So, the runner also hit the context limit.
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8386/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8386/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2698
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2698/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2698/comments
https://api.github.com/repos/ollama/ollama/issues/2698/events
https://github.com/ollama/ollama/issues/2698
2,150,287,605
I_kwDOJ0Z1Ps6AKsj1
2,698
Piping to `stdin` does not work in windows
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-02-23T03:04:37
2024-03-14T18:55:20
2024-03-14T18:55:20
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Minor issue, but piping to stdin doesn't work on windows with git bash ``` $ cat README.md | ollama run gemma "What is in this document?" failed to get console mode for stdin: The handle is invalid. ```
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2698/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2698/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2894
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2894/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2894/comments
https://api.github.com/repos/ollama/ollama/issues/2894/events
https://github.com/ollama/ollama/issues/2894
2,165,388,050
I_kwDOJ0Z1Ps6BETMS
2,894
How to get Ollama to use my RTX 4090 on windows 11
{ "login": "TimmekHW", "id": 94626112, "node_id": "U_kgDOBaPhQA", "avatar_url": "https://avatars.githubusercontent.com/u/94626112?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TimmekHW", "html_url": "https://github.com/TimmekHW", "followers_url": "https://api.github.com/users/TimmekHW/followers", "following_url": "https://api.github.com/users/TimmekHW/following{/other_user}", "gists_url": "https://api.github.com/users/TimmekHW/gists{/gist_id}", "starred_url": "https://api.github.com/users/TimmekHW/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TimmekHW/subscriptions", "organizations_url": "https://api.github.com/users/TimmekHW/orgs", "repos_url": "https://api.github.com/users/TimmekHW/repos", "events_url": "https://api.github.com/users/TimmekHW/events{/privacy}", "received_events_url": "https://api.github.com/users/TimmekHW/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-03-03T14:36:27
2024-03-03T19:19:12
2024-03-03T19:19:12
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have 12600K + 64GB RAM + RTX 4090. I use Ollama + OpenCHat. For some reason Ollama won't use my RTX 4090. How can I show the program my graphics card? ![image](https://github.com/ollama/ollama/assets/94626112/7fe5afe3-1fbb-46f1-a9e4-a1f8a58a6d05) ``` messages = chat_histories[chat_id] options = { "num_ctx": 12768, "num_thread": 10, "num_predict": 300, "repeat_last_n": 100, "temperature": 0.75 } response = "" ``` [config.json](https://github.com/ollama/ollama/files/14473300/config.json) [server.log](https://github.com/ollama/ollama/files/14473301/server.log) [app.log](https://github.com/ollama/ollama/files/14473302/app.log) **small bursts of GPU work are the work of Stable Diffusion
{ "login": "TimmekHW", "id": 94626112, "node_id": "U_kgDOBaPhQA", "avatar_url": "https://avatars.githubusercontent.com/u/94626112?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TimmekHW", "html_url": "https://github.com/TimmekHW", "followers_url": "https://api.github.com/users/TimmekHW/followers", "following_url": "https://api.github.com/users/TimmekHW/following{/other_user}", "gists_url": "https://api.github.com/users/TimmekHW/gists{/gist_id}", "starred_url": "https://api.github.com/users/TimmekHW/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/TimmekHW/subscriptions", "organizations_url": "https://api.github.com/users/TimmekHW/orgs", "repos_url": "https://api.github.com/users/TimmekHW/repos", "events_url": "https://api.github.com/users/TimmekHW/events{/privacy}", "received_events_url": "https://api.github.com/users/TimmekHW/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2894/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2894/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6394
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6394/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6394/comments
https://api.github.com/repos/ollama/ollama/issues/6394/events
https://github.com/ollama/ollama/issues/6394
2,470,933,671
I_kwDOJ0Z1Ps6TR3Sn
6,394
mistral-nemo:12b-instruct-2407-fp16 will return empty string using json mode while mistral-nemo:12b will return code
{ "login": "franz101", "id": 18228395, "node_id": "MDQ6VXNlcjE4MjI4Mzk1", "avatar_url": "https://avatars.githubusercontent.com/u/18228395?v=4", "gravatar_id": "", "url": "https://api.github.com/users/franz101", "html_url": "https://github.com/franz101", "followers_url": "https://api.github.com/users/franz101/followers", "following_url": "https://api.github.com/users/franz101/following{/other_user}", "gists_url": "https://api.github.com/users/franz101/gists{/gist_id}", "starred_url": "https://api.github.com/users/franz101/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/franz101/subscriptions", "organizations_url": "https://api.github.com/users/franz101/orgs", "repos_url": "https://api.github.com/users/franz101/repos", "events_url": "https://api.github.com/users/franz101/events{/privacy}", "received_events_url": "https://api.github.com/users/franz101/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
10
2024-08-16T20:22:03
2024-08-17T22:00:18
2024-08-17T21:59:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
currently using openai api support `mistral-nemo:12b-instruct-2407-fp16` returns an empty string
{ "login": "franz101", "id": 18228395, "node_id": "MDQ6VXNlcjE4MjI4Mzk1", "avatar_url": "https://avatars.githubusercontent.com/u/18228395?v=4", "gravatar_id": "", "url": "https://api.github.com/users/franz101", "html_url": "https://github.com/franz101", "followers_url": "https://api.github.com/users/franz101/followers", "following_url": "https://api.github.com/users/franz101/following{/other_user}", "gists_url": "https://api.github.com/users/franz101/gists{/gist_id}", "starred_url": "https://api.github.com/users/franz101/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/franz101/subscriptions", "organizations_url": "https://api.github.com/users/franz101/orgs", "repos_url": "https://api.github.com/users/franz101/repos", "events_url": "https://api.github.com/users/franz101/events{/privacy}", "received_events_url": "https://api.github.com/users/franz101/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6394/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6394/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7621
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7621/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7621/comments
https://api.github.com/repos/ollama/ollama/issues/7621/events
https://github.com/ollama/ollama/issues/7621
2,649,575,190
I_kwDOJ0Z1Ps6d7U8W
7,621
ollama run connect to server failed use 780M iGPU after update rocm-core from 6.0.2 to 6.2.2 on arch linux.
{ "login": "zw963", "id": 549126, "node_id": "MDQ6VXNlcjU0OTEyNg==", "avatar_url": "https://avatars.githubusercontent.com/u/549126?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zw963", "html_url": "https://github.com/zw963", "followers_url": "https://api.github.com/users/zw963/followers", "following_url": "https://api.github.com/users/zw963/following{/other_user}", "gists_url": "https://api.github.com/users/zw963/gists{/gist_id}", "starred_url": "https://api.github.com/users/zw963/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zw963/subscriptions", "organizations_url": "https://api.github.com/users/zw963/orgs", "repos_url": "https://api.github.com/users/zw963/repos", "events_url": "https://api.github.com/users/zw963/events{/privacy}", "received_events_url": "https://api.github.com/users/zw963/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-11-11T14:51:56
2024-11-11T15:14:30
2024-11-11T15:14:29
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I use Arch linux, i update my package to latest today. Following is my upgraded version: ``` [2024-11-11T17:25:48+0800] [ALPM] upgraded rocm-opencl-sdk (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:48+0800] [ALPM] upgraded python-pytorch-rocm (2.3.1-8 -> 2.5.1-3) [2024-11-11T17:25:45+0800] [ALPM] upgraded rocm-hip-sdk (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:45+0800] [ALPM] upgraded rocm-hip-libraries (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:44+0800] [ALPM] upgraded rocm-smi-lib (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:44+0800] [ALPM] upgraded rocm-hip-runtime (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:44+0800] [ALPM] upgraded rocm-cmake (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:44+0800] [ALPM] upgraded rocm-language-runtime (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:41+0800] [ALPM] upgraded rocm-clang-ocl (6.0.2-1 -> 6.1.2-1) [2024-11-11T17:25:41+0800] [ALPM] upgraded rocm-opencl-runtime (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:06+0800] [ALPM] upgraded rocminfo (6.0.2-1 -> 6.2.2-1) [2024-11-11T17:25:05+0800] [ALPM] upgraded rocm-device-libs (6.0.2-1 -> 6.2.2-2) [2024-11-11T17:25:05+0800] [ALPM] upgraded rocm-llvm (6.0.2-1 -> 6.2.2-2) [2024-11-11T17:24:59+0800] [ALPM] upgraded rocm-core (6.0.2-2 -> 6.2.2-1) ``` Before this update, when i run ollama serve use following command: ``` HSA_OVERRIDE_GFX_VERSION=11.0.0 OLLAMA_KEEP_ALIVE=-1 ollama serve ``` I can run `ollama run llama3.2` succssful. But, after this update, i can still run serve without issue: (check following start log) <details> <summary> ollama serve start log</summary> ``` ╰──➤ $ HSA_OVERRIDE_GFX_VERSION=11.0.0 OLLAMA_KEEP_ALIVE=-1 ollama serve 2024/11/11 22:46:36 routes.go:1189: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION:11.0.0 HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:2562047h47m16.854775807s OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/home/zw963/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR: ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2024-11-11T22:46:36.620+08:00 level=INFO source=images.go:755 msg="total blobs: 15" time=2024-11-11T22:46:36.620+08:00 level=INFO source=images.go:762 msg="total unused blobs removed: 0" time=2024-11-11T22:46:36.620+08:00 level=INFO source=routes.go:1240 msg="Listening on 127.0.0.1:11434 (version 0.4.1)" time=2024-11-11T22:46:36.621+08:00 level=INFO source=common.go:135 msg="extracting embedded files" dir=/tmp/ollama3370373680/runners time=2024-11-11T22:46:36.670+08:00 level=INFO source=common.go:49 msg="Dynamic LLM libraries" runners="[cuda_v11 cuda_v12 rocm cpu cpu_avx cpu_avx2]" time=2024-11-11T22:46:36.670+08:00 level=INFO source=gpu.go:221 msg="looking for compatible GPUs" time=2024-11-11T22:46:36.695+08:00 level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory" time=2024-11-11T22:46:36.695+08:00 level=INFO source=amd_linux.go:386 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION=11.0.0 time=2024-11-11T22:46:36.695+08:00 level=INFO source=types.go:123 msg="inference compute" id=0 library=rocm variant="" compute=gfx1103 driver=0.0 name=1002:15bf total="8.0 GiB" available="5.9 GiB" ``` </details> But, it failed with following log when run `ollama run llama3.2` ``` 2024/11/11 - 22:50:32 | 200 | 17.033µs | 127.0.0.1 | HEAD "/" [GIN] 2024/11/11 - 22:50:32 | 200 | 13.61116ms | 127.0.0.1 | POST "/api/show" time=2024-11-11T22:50:32.351+08:00 level=INFO source=sched.go:714 msg="new model will fit in available VRAM in single GPU, loading" model=/home/zw963/.ollama/models/blobs/sha256-dde5aa3fc5ffc17176b5e8bdc82f587b24b2678c6c66101bf7da77af9f7ccdff gpu=0 parallel=4 available=6257360896 required="3.7 GiB" time=2024-11-11T22:50:32.352+08:00 level=INFO source=server.go:105 msg="system memory" total="54.7 GiB" free="45.2 GiB" free_swap="63.0 GiB" time=2024-11-11T22:50:32.352+08:00 level=INFO source=memory.go:343 msg="offload to rocm" layers.requested=-1 layers.model=29 layers.offload=29 layers.split="" memory.available="[5.8 GiB]" memory.gpu_overhead="0 B" memory.required.full="3.7 GiB" memory.required.partial="3.7 GiB" memory.required.kv="896.0 MiB" memory.required.allocations="[3.7 GiB]" memory.weights.total="2.4 GiB" memory.weights.repeating="2.1 GiB" memory.weights.nonrepeating="308.2 MiB" memory.graph.full="424.0 MiB" memory.graph.partial="570.7 MiB" time=2024-11-11T22:50:32.354+08:00 level=INFO source=server.go:383 msg="starting llama server" cmd="/tmp/ollama3370373680/runners/rocm/ollama_llama_server --model /home/zw963/.ollama/models/blobs/sha256-dde5aa3fc5ffc17176b5e8bdc82f587b24b2678c6c66101bf7da77af9f7ccdff --ctx-size 8192 --batch-size 512 --n-gpu-layers 29 --threads 8 --parallel 4 --port 28751" time=2024-11-11T22:50:32.354+08:00 level=INFO source=sched.go:449 msg="loaded runners" count=1 time=2024-11-11T22:50:32.354+08:00 level=INFO source=server.go:562 msg="waiting for llama runner to start responding" time=2024-11-11T22:50:32.354+08:00 level=INFO source=server.go:596 msg="waiting for server to become available" status="llm server error" /tmp/ollama3370373680/runners/rocm/ollama_llama_server: error while loading shared libraries: libggml_rocm.so: cannot open shared object file: No such file or directory time=2024-11-11T22:50:32.605+08:00 level=ERROR source=sched.go:455 msg="error loading llama server" error="llama runner process has terminated: exit status 127" [GIN] 2024/11/11 - 22:50:32 | 500 | 288.728002ms | 127.0.0.1 | POST "/api/generate" ``` I can still use it if not set `HSA_OVERRIDE_GFX_VERSION=11.0.0` when start ollama server. ### OS Linux ### GPU AMD ### CPU AMD ### Ollama version 0.4.1
{ "login": "zw963", "id": 549126, "node_id": "MDQ6VXNlcjU0OTEyNg==", "avatar_url": "https://avatars.githubusercontent.com/u/549126?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zw963", "html_url": "https://github.com/zw963", "followers_url": "https://api.github.com/users/zw963/followers", "following_url": "https://api.github.com/users/zw963/following{/other_user}", "gists_url": "https://api.github.com/users/zw963/gists{/gist_id}", "starred_url": "https://api.github.com/users/zw963/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zw963/subscriptions", "organizations_url": "https://api.github.com/users/zw963/orgs", "repos_url": "https://api.github.com/users/zw963/repos", "events_url": "https://api.github.com/users/zw963/events{/privacy}", "received_events_url": "https://api.github.com/users/zw963/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7621/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7621/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5214
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5214/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5214/comments
https://api.github.com/repos/ollama/ollama/issues/5214/events
https://github.com/ollama/ollama/pull/5214
2,367,868,136
PR_kwDOJ0Z1Ps5zQH94
5,214
Update README.md
{ "login": "rapidarchitect", "id": 126218667, "node_id": "U_kgDOB4Xxqw", "avatar_url": "https://avatars.githubusercontent.com/u/126218667?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rapidarchitect", "html_url": "https://github.com/rapidarchitect", "followers_url": "https://api.github.com/users/rapidarchitect/followers", "following_url": "https://api.github.com/users/rapidarchitect/following{/other_user}", "gists_url": "https://api.github.com/users/rapidarchitect/gists{/gist_id}", "starred_url": "https://api.github.com/users/rapidarchitect/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rapidarchitect/subscriptions", "organizations_url": "https://api.github.com/users/rapidarchitect/orgs", "repos_url": "https://api.github.com/users/rapidarchitect/repos", "events_url": "https://api.github.com/users/rapidarchitect/events{/privacy}", "received_events_url": "https://api.github.com/users/rapidarchitect/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-22T15:08:49
2024-07-01T02:00:58
2024-07-01T02:00:58
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5214", "html_url": "https://github.com/ollama/ollama/pull/5214", "diff_url": "https://github.com/ollama/ollama/pull/5214.diff", "patch_url": "https://github.com/ollama/ollama/pull/5214.patch", "merged_at": "2024-07-01T02:00:58" }
Added Mesop example to web & desktop
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5214/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5214/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1780
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1780/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1780/comments
https://api.github.com/repos/ollama/ollama/issues/1780/events
https://github.com/ollama/ollama/pull/1780
2,064,804,933
PR_kwDOJ0Z1Ps5jLeS2
1,780
update cmake flags for `amd64` macOS
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-01-04T00:06:12
2024-01-04T00:22:16
2024-01-04T00:22:15
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1780", "html_url": "https://github.com/ollama/ollama/pull/1780", "diff_url": "https://github.com/ollama/ollama/pull/1780.diff", "patch_url": "https://github.com/ollama/ollama/pull/1780.patch", "merged_at": "2024-01-04T00:22:15" }
null
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1780/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1780/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4832
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4832/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4832/comments
https://api.github.com/repos/ollama/ollama/issues/4832/events
https://github.com/ollama/ollama/issues/4832
2,335,432,838
I_kwDOJ0Z1Ps6LM-CG
4,832
llama3:7b cache size set
{ "login": "ciscoivan", "id": 55469637, "node_id": "MDQ6VXNlcjU1NDY5NjM3", "avatar_url": "https://avatars.githubusercontent.com/u/55469637?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ciscoivan", "html_url": "https://github.com/ciscoivan", "followers_url": "https://api.github.com/users/ciscoivan/followers", "following_url": "https://api.github.com/users/ciscoivan/following{/other_user}", "gists_url": "https://api.github.com/users/ciscoivan/gists{/gist_id}", "starred_url": "https://api.github.com/users/ciscoivan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ciscoivan/subscriptions", "organizations_url": "https://api.github.com/users/ciscoivan/orgs", "repos_url": "https://api.github.com/users/ciscoivan/repos", "events_url": "https://api.github.com/users/ciscoivan/events{/privacy}", "received_events_url": "https://api.github.com/users/ciscoivan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-06-05T09:53:43
2024-06-09T17:35:53
2024-06-09T17:35:53
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
![1](https://github.com/ollama/ollama/assets/55469637/440bb5f3-d605-4962-a895-3205d7c9d621) I installed two NVIDIA RTX 2080 TI graphics cards in an experimental deployment and successfully ran the llama3:7b model. I want to know how to adjust the cache size. thanks
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4832/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4832/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8614
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8614/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8614/comments
https://api.github.com/repos/ollama/ollama/issues/8614/events
https://github.com/ollama/ollama/issues/8614
2,813,943,892
I_kwDOJ0Z1Ps6nuWBU
8,614
Problems with deepseek-r1:671b, ollama keeps crashing on long answers
{ "login": "fabiounixpi", "id": 48057600, "node_id": "MDQ6VXNlcjQ4MDU3NjAw", "avatar_url": "https://avatars.githubusercontent.com/u/48057600?v=4", "gravatar_id": "", "url": "https://api.github.com/users/fabiounixpi", "html_url": "https://github.com/fabiounixpi", "followers_url": "https://api.github.com/users/fabiounixpi/followers", "following_url": "https://api.github.com/users/fabiounixpi/following{/other_user}", "gists_url": "https://api.github.com/users/fabiounixpi/gists{/gist_id}", "starred_url": "https://api.github.com/users/fabiounixpi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/fabiounixpi/subscriptions", "organizations_url": "https://api.github.com/users/fabiounixpi/orgs", "repos_url": "https://api.github.com/users/fabiounixpi/repos", "events_url": "https://api.github.com/users/fabiounixpi/events{/privacy}", "received_events_url": "https://api.github.com/users/fabiounixpi/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
11
2025-01-27T20:04:40
2025-01-30T13:07:47
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Hi all, I'm using an r960 with 2TB of ram, so ram is not a problem here. I'm experiencing constant crashes of ollama 0.5.7 and deepseek-r1:671b, even increasing the context window with the command /set parameter num_ctx 4096. I also tried a second system, an r670 csp with 1TB of ram, but the problem occurs in the same way. I'm not able to use gpu due to the massive size of the model, anyway plenty of cores do the job for my current pourposes. os are ubuntu 22.04.5 and 24.04.1 ### OS Linux ### GPU _No response_ ### CPU Intel ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8614/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8614/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5309
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5309/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5309/comments
https://api.github.com/repos/ollama/ollama/issues/5309/events
https://github.com/ollama/ollama/pull/5309
2,376,190,397
PR_kwDOJ0Z1Ps5zrkYW
5,309
Update OpenAI Compatibility Docs with /v1/models/{model}
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-06-26T20:17:09
2024-08-01T23:00:44
2024-08-01T22:58:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5309", "html_url": "https://github.com/ollama/ollama/pull/5309", "diff_url": "https://github.com/ollama/ollama/pull/5309.diff", "patch_url": "https://github.com/ollama/ollama/pull/5309.patch", "merged_at": "2024-08-01T22:58:13" }
null
{ "login": "royjhan", "id": 65097070, "node_id": "MDQ6VXNlcjY1MDk3MDcw", "avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4", "gravatar_id": "", "url": "https://api.github.com/users/royjhan", "html_url": "https://github.com/royjhan", "followers_url": "https://api.github.com/users/royjhan/followers", "following_url": "https://api.github.com/users/royjhan/following{/other_user}", "gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}", "starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/royjhan/subscriptions", "organizations_url": "https://api.github.com/users/royjhan/orgs", "repos_url": "https://api.github.com/users/royjhan/repos", "events_url": "https://api.github.com/users/royjhan/events{/privacy}", "received_events_url": "https://api.github.com/users/royjhan/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5309/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5309/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/72
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/72/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/72/comments
https://api.github.com/repos/ollama/ollama/issues/72/events
https://github.com/ollama/ollama/issues/72
1,800,080,847
I_kwDOJ0Z1Ps5rSw3P
72
`ollama run` doesn't continue after one reponse
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2023-07-12T03:06:00
2023-07-17T16:43:08
2023-07-17T16:43:08
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
here are how you reproduce ```$ ollama run orca hello Hello! How can I assist you today?Error: stream: EOF $ logs ollama run orca "why is the sky blue" The sky appears blue because of a process called scattering. When sunlight enters the Earth's atmosphere, it collides with gas molecules such as oxygen and nitrogen. These collisions cause the light to scatter in all directions. Blue light has a shorter wavelength and is scattered more easily than other colors, so it is scattered more widely across the sky, making it appear blue. This effect is also why the sky is usually darker during sunrise and sunset when the sun is below the horizon and cannot be seen.Error: stream: EOF```
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/72/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/72/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1679
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1679/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1679/comments
https://api.github.com/repos/ollama/ollama/issues/1679/events
https://github.com/ollama/ollama/pull/1679
2,054,480,828
PR_kwDOJ0Z1Ps5irqMT
1,679
build cuda and rocm
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-12-22T20:22:28
2024-01-26T00:38:15
2024-01-26T00:38:14
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1679", "html_url": "https://github.com/ollama/ollama/pull/1679", "diff_url": "https://github.com/ollama/ollama/pull/1679.diff", "patch_url": "https://github.com/ollama/ollama/pull/1679.patch", "merged_at": "2024-01-26T00:38:14" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1679/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1679/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2559
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2559/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2559/comments
https://api.github.com/repos/ollama/ollama/issues/2559/events
https://github.com/ollama/ollama/issues/2559
2,139,912,869
I_kwDOJ0Z1Ps5_jHql
2,559
Feature - Support Custom Actions
{ "login": "joeldhenry", "id": 12555860, "node_id": "MDQ6VXNlcjEyNTU1ODYw", "avatar_url": "https://avatars.githubusercontent.com/u/12555860?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeldhenry", "html_url": "https://github.com/joeldhenry", "followers_url": "https://api.github.com/users/joeldhenry/followers", "following_url": "https://api.github.com/users/joeldhenry/following{/other_user}", "gists_url": "https://api.github.com/users/joeldhenry/gists{/gist_id}", "starred_url": "https://api.github.com/users/joeldhenry/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joeldhenry/subscriptions", "organizations_url": "https://api.github.com/users/joeldhenry/orgs", "repos_url": "https://api.github.com/users/joeldhenry/repos", "events_url": "https://api.github.com/users/joeldhenry/events{/privacy}", "received_events_url": "https://api.github.com/users/joeldhenry/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q", "url": "https://api.github.com/repos/ollama/ollama/labels/api", "name": "api", "color": "bfdadc", "default": false, "description": "" } ]
open
false
null
[]
null
1
2024-02-17T09:19:23
2024-11-06T18:56:34
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Support for custom actions to cal custom API/Code as part of llama response. possibly part of modelfiles via python scripts? chatGPT has similar with integrations with Zapier
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2559/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2559/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5238
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5238/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5238/comments
https://api.github.com/repos/ollama/ollama/issues/5238/events
https://github.com/ollama/ollama/issues/5238
2,368,588,675
I_kwDOJ0Z1Ps6NLcuD
5,238
How to update Ollama to the latest version?
{ "login": "qzc438", "id": 61488260, "node_id": "MDQ6VXNlcjYxNDg4MjYw", "avatar_url": "https://avatars.githubusercontent.com/u/61488260?v=4", "gravatar_id": "", "url": "https://api.github.com/users/qzc438", "html_url": "https://github.com/qzc438", "followers_url": "https://api.github.com/users/qzc438/followers", "following_url": "https://api.github.com/users/qzc438/following{/other_user}", "gists_url": "https://api.github.com/users/qzc438/gists{/gist_id}", "starred_url": "https://api.github.com/users/qzc438/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qzc438/subscriptions", "organizations_url": "https://api.github.com/users/qzc438/orgs", "repos_url": "https://api.github.com/users/qzc438/repos", "events_url": "https://api.github.com/users/qzc438/events{/privacy}", "received_events_url": "https://api.github.com/users/qzc438/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2024-06-23T14:24:45
2024-06-24T11:19:25
2024-06-24T11:19:25
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? As the title described. How to update Ollama to the latest version? ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
{ "login": "qzc438", "id": 61488260, "node_id": "MDQ6VXNlcjYxNDg4MjYw", "avatar_url": "https://avatars.githubusercontent.com/u/61488260?v=4", "gravatar_id": "", "url": "https://api.github.com/users/qzc438", "html_url": "https://github.com/qzc438", "followers_url": "https://api.github.com/users/qzc438/followers", "following_url": "https://api.github.com/users/qzc438/following{/other_user}", "gists_url": "https://api.github.com/users/qzc438/gists{/gist_id}", "starred_url": "https://api.github.com/users/qzc438/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/qzc438/subscriptions", "organizations_url": "https://api.github.com/users/qzc438/orgs", "repos_url": "https://api.github.com/users/qzc438/repos", "events_url": "https://api.github.com/users/qzc438/events{/privacy}", "received_events_url": "https://api.github.com/users/qzc438/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5238/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5238/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5242
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5242/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5242/comments
https://api.github.com/repos/ollama/ollama/issues/5242/events
https://github.com/ollama/ollama/issues/5242
2,368,823,955
I_kwDOJ0Z1Ps6NMWKT
5,242
Slow performance on `/api/show`
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2024-06-23T18:36:53
2024-07-24T19:10:57
2024-07-24T19:10:57
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Because we now show more model details, `/api/show` has gotten slower. The part that's slow specifically is reading the arrays (vocab, tensors, etc) ``` case ggufTypeArray: v, err = readGGUFArray(llm, rs) ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version 0.1.45
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5242/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5242/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3512
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3512/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3512/comments
https://api.github.com/repos/ollama/ollama/issues/3512/events
https://github.com/ollama/ollama/issues/3512
2,229,136,933
I_kwDOJ0Z1Ps6E3e4l
3,512
Experimental LLM Library Override does not appear to work on Windows
{ "login": "lrq3000", "id": 1118942, "node_id": "MDQ6VXNlcjExMTg5NDI=", "avatar_url": "https://avatars.githubusercontent.com/u/1118942?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lrq3000", "html_url": "https://github.com/lrq3000", "followers_url": "https://api.github.com/users/lrq3000/followers", "following_url": "https://api.github.com/users/lrq3000/following{/other_user}", "gists_url": "https://api.github.com/users/lrq3000/gists{/gist_id}", "starred_url": "https://api.github.com/users/lrq3000/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lrq3000/subscriptions", "organizations_url": "https://api.github.com/users/lrq3000/orgs", "repos_url": "https://api.github.com/users/lrq3000/repos", "events_url": "https://api.github.com/users/lrq3000/events{/privacy}", "received_events_url": "https://api.github.com/users/lrq3000/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
4
2024-04-06T08:14:22
2024-04-23T19:40:16
2024-04-23T02:06:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I tried the [Experimental LLM Library Override](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#llm-libraries) on Windows via two means: * Temporary environment variable definition: `SET OLLAMA_LLM_LIBRARY="cpu_avx2" & ollama run deepseek-coder` * Permanent environment variable definition in the Windows System dialog. Both failed, in server.log I get the following error: `time=2024-04-06T10:11:46.333+02:00 level=INFO source=llm.go:147 msg="Invalid OLLAMA_LLM_LIBRARY \"cpu_avx2\" - not found"` And ollama proceeds to use my GPU. See the full server.log attached: [server.log](https://github.com/ollama/ollama/files/14892936/server.log) ### What did you expect to see? Ollama should be using cpu_avx2 instead of the GPU. ### Steps to reproduce _No response_ ### Are there any recent changes that introduced the issue? _No response_ ### OS Windows ### Architecture x86 ### Platform _No response_ ### Ollama version 0.1.30 ### GPU Nvidia ### GPU info Nvidia GeForce 3060 Laptop ### CPU Intel ### Other software Intel i7-12700H
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3512/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3512/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4497
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4497/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4497/comments
https://api.github.com/repos/ollama/ollama/issues/4497/events
https://github.com/ollama/ollama/issues/4497
2,302,421,423
I_kwDOJ0Z1Ps6JPCmv
4,497
Ollama 0.1.38 has high video memory usage and runs very slowly.
{ "login": "chenwei0930", "id": 17743683, "node_id": "MDQ6VXNlcjE3NzQzNjgz", "avatar_url": "https://avatars.githubusercontent.com/u/17743683?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chenwei0930", "html_url": "https://github.com/chenwei0930", "followers_url": "https://api.github.com/users/chenwei0930/followers", "following_url": "https://api.github.com/users/chenwei0930/following{/other_user}", "gists_url": "https://api.github.com/users/chenwei0930/gists{/gist_id}", "starred_url": "https://api.github.com/users/chenwei0930/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chenwei0930/subscriptions", "organizations_url": "https://api.github.com/users/chenwei0930/orgs", "repos_url": "https://api.github.com/users/chenwei0930/repos", "events_url": "https://api.github.com/users/chenwei0930/events{/privacy}", "received_events_url": "https://api.github.com/users/chenwei0930/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
4
2024-05-17T10:54:55
2024-06-22T07:07:40
2024-06-21T23:34:11
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am using Windows 10 with an NVIDIA 2080Ti graphics card that has 22GB of video memory. I upgraded from version 0.1.32 to 0.1.38 with the goal of supporting loading multiple models and handling multiple concurrent requests. However, I noticed that under version 0.1.38, the video memory usage is very high, and the speed has become much slower. I am using the "codeqwen:7b-chat-v1.5-q8_0" model. Under version 0.1.32, it used around 8GB of video memory and output approximately 10 tokens per second. However, under version 0.1.38, it is using 18.8GB of video memory, and based on my observation, it is only outputting 1-2 tokens per second. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.38
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4497/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4497/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8190
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8190/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8190/comments
https://api.github.com/repos/ollama/ollama/issues/8190/events
https://github.com/ollama/ollama/pull/8190
2,753,561,133
PR_kwDOJ0Z1Ps6F9VTg
8,190
macos: detect potential version skew
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2024-12-20T22:17:24
2024-12-23T15:35:57
null
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8190", "html_url": "https://github.com/ollama/ollama/pull/8190", "diff_url": "https://github.com/ollama/ollama/pull/8190.diff", "patch_url": "https://github.com/ollama/ollama/pull/8190.patch", "merged_at": null }
During upgrade, we could get into a scenario where an old serve tries to start newer runner executables, which could have new expectations. This added check will handle non-zero exit status from the runner and double check the current process has the same version as the executable on disk. If the version has skewed, exit the serve process so the app can respawn the new binary.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8190/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8190/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6134
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6134/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6134/comments
https://api.github.com/repos/ollama/ollama/issues/6134/events
https://github.com/ollama/ollama/issues/6134
2,443,859,187
I_kwDOJ0Z1Ps6RqlTz
6,134
can't change ollama server address :127.0.0.1:11434 after binary ollama install
{ "login": "cnopens", "id": 3257702, "node_id": "MDQ6VXNlcjMyNTc3MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/3257702?v=4", "gravatar_id": "", "url": "https://api.github.com/users/cnopens", "html_url": "https://github.com/cnopens", "followers_url": "https://api.github.com/users/cnopens/followers", "following_url": "https://api.github.com/users/cnopens/following{/other_user}", "gists_url": "https://api.github.com/users/cnopens/gists{/gist_id}", "starred_url": "https://api.github.com/users/cnopens/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cnopens/subscriptions", "organizations_url": "https://api.github.com/users/cnopens/orgs", "repos_url": "https://api.github.com/users/cnopens/repos", "events_url": "https://api.github.com/users/cnopens/events{/privacy}", "received_events_url": "https://api.github.com/users/cnopens/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-08-02T02:23:32
2024-08-09T21:13:42
2024-08-09T21:13:29
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? i install ollama using binary package ,find that ollama server address 127.0.0.1:11434 ,ip why can't change it ? who run into the problem ? later i only used nginx proxy ,but ,open-webui reponse-webui was very slow vs command consle ### OS Linux ### GPU Intel ### CPU Intel ### Ollama version 0.3.2
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6134/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6134/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1423
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1423/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1423/comments
https://api.github.com/repos/ollama/ollama/issues/1423/events
https://github.com/ollama/ollama/issues/1423
2,031,667,493
I_kwDOJ0Z1Ps55GMkl
1,423
Allow Response Templating
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2023-12-07T23:11:26
2023-12-22T22:07:06
2023-12-22T22:07:06
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
In order to support formats like chatml Ollama must support post-response templating: ``` <|im_start|>user Hi there!<|im_end|> <|im_start|>assistant Nice to meet you!<|im_end|> ```
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1423/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1423/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7419
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7419/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7419/comments
https://api.github.com/repos/ollama/ollama/issues/7419/events
https://github.com/ollama/ollama/issues/7419
2,624,183,476
I_kwDOJ0Z1Ps6cady0
7,419
Integrating Into Desktop App
{ "login": "brian-at-pieces", "id": 98757707, "node_id": "U_kgDOBeLsSw", "avatar_url": "https://avatars.githubusercontent.com/u/98757707?v=4", "gravatar_id": "", "url": "https://api.github.com/users/brian-at-pieces", "html_url": "https://github.com/brian-at-pieces", "followers_url": "https://api.github.com/users/brian-at-pieces/followers", "following_url": "https://api.github.com/users/brian-at-pieces/following{/other_user}", "gists_url": "https://api.github.com/users/brian-at-pieces/gists{/gist_id}", "starred_url": "https://api.github.com/users/brian-at-pieces/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/brian-at-pieces/subscriptions", "organizations_url": "https://api.github.com/users/brian-at-pieces/orgs", "repos_url": "https://api.github.com/users/brian-at-pieces/repos", "events_url": "https://api.github.com/users/brian-at-pieces/events{/privacy}", "received_events_url": "https://api.github.com/users/brian-at-pieces/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-10-30T13:52:19
2024-11-01T18:56:26
2024-11-01T18:56:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'd love to use Ollama for serving LLMs in my company's Mac/Linux/Windows desktop app, but I'm a little confused about some things. I'd like to integrate it directly rather than requring the user to manually install Ollama because that UX isn't very good IMO. You [mention in the Windows docs](https://github.com/ollama/ollama/blob/main/docs/windows.md#standalone-cli:~:text=If%20you%27d%20like%20to%20install%20or%20integrate%20Ollama%20as%20a%20service%2C%20a%20standalone%20ollama%2Dwindows%2Damd64.zip%20zip%20file%20is%20available%20containing%20only%20the%20Ollama%20CLI%20and%20GPU%20library%20dependencies%20for%20Nvidia%20and%20AMD.%20This%20allows%20for%20embedding%20Ollama%20in%20existing%20applications) that this can be done by just utilizing the Windows release zip, but at 3.25 GB that's way too large to bundle with our app. It looks like it's so big because (1) different runners are included to allow Ollama to choose the optimal one, and (2) AMD and NVIDIA GPU libs are included. I understand including different runners, but I don't understand why GPU libs like cudart, cublas, rocblas, etc. are included - shouldn't these be present on the system if the proper drivers are installed? If it's not possible to remove these, then it would be nice if you could generate separate releases for each runner / gpu libs to somewhat reduce build size. That being said, thanks for making an awesome product!
{ "login": "brian-at-pieces", "id": 98757707, "node_id": "U_kgDOBeLsSw", "avatar_url": "https://avatars.githubusercontent.com/u/98757707?v=4", "gravatar_id": "", "url": "https://api.github.com/users/brian-at-pieces", "html_url": "https://github.com/brian-at-pieces", "followers_url": "https://api.github.com/users/brian-at-pieces/followers", "following_url": "https://api.github.com/users/brian-at-pieces/following{/other_user}", "gists_url": "https://api.github.com/users/brian-at-pieces/gists{/gist_id}", "starred_url": "https://api.github.com/users/brian-at-pieces/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/brian-at-pieces/subscriptions", "organizations_url": "https://api.github.com/users/brian-at-pieces/orgs", "repos_url": "https://api.github.com/users/brian-at-pieces/repos", "events_url": "https://api.github.com/users/brian-at-pieces/events{/privacy}", "received_events_url": "https://api.github.com/users/brian-at-pieces/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7419/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7419/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8556
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8556/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8556/comments
https://api.github.com/repos/ollama/ollama/issues/8556/events
https://github.com/ollama/ollama/issues/8556
2,808,386,790
I_kwDOJ0Z1Ps6nZJTm
8,556
Please separate deepseek-r1 from deepseek-r1-Distill!
{ "login": "win10ogod", "id": 125795763, "node_id": "U_kgDOB399sw", "avatar_url": "https://avatars.githubusercontent.com/u/125795763?v=4", "gravatar_id": "", "url": "https://api.github.com/users/win10ogod", "html_url": "https://github.com/win10ogod", "followers_url": "https://api.github.com/users/win10ogod/followers", "following_url": "https://api.github.com/users/win10ogod/following{/other_user}", "gists_url": "https://api.github.com/users/win10ogod/gists{/gist_id}", "starred_url": "https://api.github.com/users/win10ogod/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/win10ogod/subscriptions", "organizations_url": "https://api.github.com/users/win10ogod/orgs", "repos_url": "https://api.github.com/users/win10ogod/repos", "events_url": "https://api.github.com/users/win10ogod/events{/privacy}", "received_events_url": "https://api.github.com/users/win10ogod/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
0
2025-01-24T03:04:56
2025-01-24T03:20:18
2025-01-24T03:20:18
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Please separate deepseek-r1 from deepseek-r1-Distill! This is not the same model and the architecture is different! The model on the ollama official website is a perfect obfuscation!
{ "login": "win10ogod", "id": 125795763, "node_id": "U_kgDOB399sw", "avatar_url": "https://avatars.githubusercontent.com/u/125795763?v=4", "gravatar_id": "", "url": "https://api.github.com/users/win10ogod", "html_url": "https://github.com/win10ogod", "followers_url": "https://api.github.com/users/win10ogod/followers", "following_url": "https://api.github.com/users/win10ogod/following{/other_user}", "gists_url": "https://api.github.com/users/win10ogod/gists{/gist_id}", "starred_url": "https://api.github.com/users/win10ogod/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/win10ogod/subscriptions", "organizations_url": "https://api.github.com/users/win10ogod/orgs", "repos_url": "https://api.github.com/users/win10ogod/repos", "events_url": "https://api.github.com/users/win10ogod/events{/privacy}", "received_events_url": "https://api.github.com/users/win10ogod/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8556/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8556/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5790
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5790/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5790/comments
https://api.github.com/repos/ollama/ollama/issues/5790/events
https://github.com/ollama/ollama/pull/5790
2,418,358,087
PR_kwDOJ0Z1Ps513q8w
5,790
Update llama.cpp submodule to 1bdd8ae1
{ "login": "zhongTao99", "id": 56594937, "node_id": "MDQ6VXNlcjU2NTk0OTM3", "avatar_url": "https://avatars.githubusercontent.com/u/56594937?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zhongTao99", "html_url": "https://github.com/zhongTao99", "followers_url": "https://api.github.com/users/zhongTao99/followers", "following_url": "https://api.github.com/users/zhongTao99/following{/other_user}", "gists_url": "https://api.github.com/users/zhongTao99/gists{/gist_id}", "starred_url": "https://api.github.com/users/zhongTao99/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zhongTao99/subscriptions", "organizations_url": "https://api.github.com/users/zhongTao99/orgs", "repos_url": "https://api.github.com/users/zhongTao99/repos", "events_url": "https://api.github.com/users/zhongTao99/events{/privacy}", "received_events_url": "https://api.github.com/users/zhongTao99/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-07-19T08:28:04
2024-09-03T17:20:14
2024-09-03T17:20:13
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5790", "html_url": "https://github.com/ollama/ollama/pull/5790", "diff_url": "https://github.com/ollama/ollama/pull/5790.diff", "patch_url": "https://github.com/ollama/ollama/pull/5790.patch", "merged_at": null }
fix:https://github.com/ollama/ollama/issues/5769
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5790/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5790/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7669
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7669/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7669/comments
https://api.github.com/repos/ollama/ollama/issues/7669/events
https://github.com/ollama/ollama/issues/7669
2,659,895,139
I_kwDOJ0Z1Ps6eisdj
7,669
Only CPU is used after rebooting
{ "login": "3DAlgoLab", "id": 83936830, "node_id": "MDQ6VXNlcjgzOTM2ODMw", "avatar_url": "https://avatars.githubusercontent.com/u/83936830?v=4", "gravatar_id": "", "url": "https://api.github.com/users/3DAlgoLab", "html_url": "https://github.com/3DAlgoLab", "followers_url": "https://api.github.com/users/3DAlgoLab/followers", "following_url": "https://api.github.com/users/3DAlgoLab/following{/other_user}", "gists_url": "https://api.github.com/users/3DAlgoLab/gists{/gist_id}", "starred_url": "https://api.github.com/users/3DAlgoLab/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/3DAlgoLab/subscriptions", "organizations_url": "https://api.github.com/users/3DAlgoLab/orgs", "repos_url": "https://api.github.com/users/3DAlgoLab/repos", "events_url": "https://api.github.com/users/3DAlgoLab/events{/privacy}", "received_events_url": "https://api.github.com/users/3DAlgoLab/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5755339642, "node_id": "LA_kwDOJ0Z1Ps8AAAABVwuDeg", "url": "https://api.github.com/repos/ollama/ollama/labels/linux", "name": "linux", "color": "516E70", "default": false, "description": "" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
8
2024-11-14T19:49:01
2024-11-19T03:42:01
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
[I found someone wrote a thread describing only cpu is used after rebooting in windows ](https://github.com/ollama/ollama/issues/4984#issue-2347076913) I also had similar problems even in Ubuntu OS. I used the latest version(0.4.1). I guess this bug comes from that the ollama service is started faster than the init. of GPUs. So I make an **ad-hoc** solution. Instead of service, I just make a script delaying start serve(ollama serve). ```bash # ollama_run echo "Delayed Ollama Runner Start, it delays 10 sec." sleep 10 ollama serve ``` Then I make this called from *Ubuntu Startup Application Preferences*. I think its delaying may not be needed as it is called after GPUs initialization is finished anyway. ![capture 2024-11-15 042959](https://github.com/user-attachments/assets/80ca322d-49ea-4baa-991d-15b1413612ef) **WARNING**: After starting ollama by calling directly `ollama serve`, model storage directory is changed to `~/.ollama/models`.(I don't know why?) So previously dowonloaded model is not loaded. In that case, you can copy or move whole models folder to '[home folder]/.ollama` from `/usr/share/ollama/.ollama`. _Originally posted by @3DAlgoLab in https://github.com/ollama/ollama/issues/4984#issuecomment-2477251430_
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7669/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7669/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/249
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/249/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/249/comments
https://api.github.com/repos/ollama/ollama/issues/249/events
https://github.com/ollama/ollama/pull/249
1,830,020,783
PR_kwDOJ0Z1Ps5W2Bla
249
Add "Awesome projects built with Ollama" section to README, including Continue
{ "login": "sestinj", "id": 33237525, "node_id": "MDQ6VXNlcjMzMjM3NTI1", "avatar_url": "https://avatars.githubusercontent.com/u/33237525?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sestinj", "html_url": "https://github.com/sestinj", "followers_url": "https://api.github.com/users/sestinj/followers", "following_url": "https://api.github.com/users/sestinj/following{/other_user}", "gists_url": "https://api.github.com/users/sestinj/gists{/gist_id}", "starred_url": "https://api.github.com/users/sestinj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sestinj/subscriptions", "organizations_url": "https://api.github.com/users/sestinj/orgs", "repos_url": "https://api.github.com/users/sestinj/repos", "events_url": "https://api.github.com/users/sestinj/events{/privacy}", "received_events_url": "https://api.github.com/users/sestinj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-07-31T21:01:22
2023-08-01T15:07:50
2023-08-01T15:07:50
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/249", "html_url": "https://github.com/ollama/ollama/pull/249", "diff_url": "https://github.com/ollama/ollama/pull/249.diff", "patch_url": "https://github.com/ollama/ollama/pull/249.patch", "merged_at": "2023-08-01T15:07:50" }
Format and text are up for debate, but here's a description of Continue
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/249/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/249/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2569
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2569/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2569/comments
https://api.github.com/repos/ollama/ollama/issues/2569/events
https://github.com/ollama/ollama/issues/2569
2,140,757,706
I_kwDOJ0Z1Ps5_mV7K
2,569
Connection with http://127.0.0.1:11434/api/chat forcibly closed
{ "login": "spampinato55", "id": 47316524, "node_id": "MDQ6VXNlcjQ3MzE2NTI0", "avatar_url": "https://avatars.githubusercontent.com/u/47316524?v=4", "gravatar_id": "", "url": "https://api.github.com/users/spampinato55", "html_url": "https://github.com/spampinato55", "followers_url": "https://api.github.com/users/spampinato55/followers", "following_url": "https://api.github.com/users/spampinato55/following{/other_user}", "gists_url": "https://api.github.com/users/spampinato55/gists{/gist_id}", "starred_url": "https://api.github.com/users/spampinato55/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/spampinato55/subscriptions", "organizations_url": "https://api.github.com/users/spampinato55/orgs", "repos_url": "https://api.github.com/users/spampinato55/repos", "events_url": "https://api.github.com/users/spampinato55/events{/privacy}", "received_events_url": "https://api.github.com/users/spampinato55/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
15
2024-02-18T05:47:39
2024-08-25T11:21:20
2024-02-20T21:56:06
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I've installed Ollama in Windows 10, I launch it and it runs, I can pull a model but when I want to run it this is the error message I see: "Error: Post "http://127.0.0.1:11434/api/chat": read tcp 127.0.0.1:52725->127.0.0.1:11434: wsarecv: An existing connection was forcibly closed by the remote host." I disabled the previous wsl service, I've also set the the port 11434 in the firewall but nothing happens. With WSL it runs. Thanks
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2569/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2569/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3085
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3085/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3085/comments
https://api.github.com/repos/ollama/ollama/issues/3085/events
https://github.com/ollama/ollama/issues/3085
2,182,679,116
I_kwDOJ0Z1Ps6CGQpM
3,085
Please support Zephyr 7B Gemma
{ "login": "RahulBhalley", "id": 9640948, "node_id": "MDQ6VXNlcjk2NDA5NDg=", "avatar_url": "https://avatars.githubusercontent.com/u/9640948?v=4", "gravatar_id": "", "url": "https://api.github.com/users/RahulBhalley", "html_url": "https://github.com/RahulBhalley", "followers_url": "https://api.github.com/users/RahulBhalley/followers", "following_url": "https://api.github.com/users/RahulBhalley/following{/other_user}", "gists_url": "https://api.github.com/users/RahulBhalley/gists{/gist_id}", "starred_url": "https://api.github.com/users/RahulBhalley/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/RahulBhalley/subscriptions", "organizations_url": "https://api.github.com/users/RahulBhalley/orgs", "repos_url": "https://api.github.com/users/RahulBhalley/repos", "events_url": "https://api.github.com/users/RahulBhalley/events{/privacy}", "received_events_url": "https://api.github.com/users/RahulBhalley/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
2
2024-03-12T21:09:21
2024-03-13T10:22:06
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Please support [Zephyr 7B Gemma](https://huggingface.co/HuggingFaceH4/zephyr-7b-gemma-v0.1)! This [HG Chat](https://huggingface.co/spaces/HuggingFaceH4/zephyr-7b-gemma-chat) is a lot better than Zephyr beta (fine-tuned on Mistral 7B).
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3085/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3085/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7072
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7072/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7072/comments
https://api.github.com/repos/ollama/ollama/issues/7072/events
https://github.com/ollama/ollama/issues/7072
2,560,411,081
I_kwDOJ0Z1Ps6YnMXJ
7,072
Deepseek-v2.5 fails to load on a system with 24GB VRAM (RTX 3090) and 128GB RAM
{ "login": "LeonidShamis", "id": 1818114, "node_id": "MDQ6VXNlcjE4MTgxMTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1818114?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LeonidShamis", "html_url": "https://github.com/LeonidShamis", "followers_url": "https://api.github.com/users/LeonidShamis/followers", "following_url": "https://api.github.com/users/LeonidShamis/following{/other_user}", "gists_url": "https://api.github.com/users/LeonidShamis/gists{/gist_id}", "starred_url": "https://api.github.com/users/LeonidShamis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LeonidShamis/subscriptions", "organizations_url": "https://api.github.com/users/LeonidShamis/orgs", "repos_url": "https://api.github.com/users/LeonidShamis/repos", "events_url": "https://api.github.com/users/LeonidShamis/events{/privacy}", "received_events_url": "https://api.github.com/users/LeonidShamis/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
4
2024-10-02T00:06:08
2024-10-27T05:25:22
2024-10-02T05:54:01
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I'm unable to load the [deepseek-v2.5](https://ollama.com/library/deepseek-v2.5) model on a system with 24GB VRAM (RTX 3090) and 128GB RAM: ``` $ ollama --version ollama version is 0.3.11 $ $ ollama list | grep -e ID -e deepseek-v2.5 NAME ID SIZE MODIFIED deepseek-v2.5:latest 409b2dd8a3c4 132 GB 9 hours ago $ ollama show deepseek-v2.5 Model architecture deepseek2 parameters 235.7B context length 163840 embedding length 5120 quantization Q4_0 Parameters stop "<|begin?of?sentence|>" stop "<|end?of?sentence|>" stop "<|User|>" stop "<|Assistant|>" stop "<|fim?begin|>" stop "<|fim?hole|>" stop "<|fim?end|>" License DEEPSEEK LICENSE AGREEMENT Version 1.0, 23 October 2023 $ ollama run deepseek-v2.5 Error: llama runner process has terminated: error:failed to create context with model '/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c' $ ``` Journatctl output: ``` $ sudo journalctl -f Oct 02 09:20:41 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:20:41 | 200 | 14.89µs | 127.0.0.1 | HEAD "/" Oct 02 09:20:41 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:20:41 | 200 | 22.898622ms | 127.0.0.1 | GET "/api/tags" Oct 02 09:20:56 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:21:04 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:21:08 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:08 | 200 | 18.64µs | 127.0.0.1 | HEAD "/" Oct 02 09:21:08 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:08 | 200 | 3.682586ms | 127.0.0.1 | GET "/api/tags" Oct 02 09:21:23 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:23 | 200 | 29.73µs | 127.0.0.1 | HEAD "/" Oct 02 09:21:23 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:23 | 200 | 3.880099ms | 127.0.0.1 | GET "/api/tags" Oct 02 09:21:28 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:28 | 200 | 15.64µs | 127.0.0.1 | HEAD "/" Oct 02 09:21:28 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:28 | 200 | 1.227682ms | 127.0.0.1 | GET "/api/tags" Oct 02 09:21:35 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:35 | 200 | 17.3µs | 127.0.0.1 | HEAD "/" Oct 02 09:21:35 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:21:35 | 200 | 29.058041ms | 127.0.0.1 | POST "/api/show" Oct 02 09:22:02 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:22:02 | 200 | 15.01µs | 127.0.0.1 | HEAD "/" Oct 02 09:22:02 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:22:02 | 200 | 13.210104ms | 127.0.0.1 | POST "/api/show" Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.094+10:00 level=INFO source=sched.go:507 msg="updated VRAM based on existing loaded models" gpu=GPU-cebf1f58-0f47-0efa-a683-ceeb3c1755cc library=cuda total="23.7 GiB" available="8.8 GiB" Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.909+10:00 level=INFO source=server.go:103 msg="system memory" total="125.7 GiB" free="118.6 GiB" free_swap="7.2 GiB" Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.910+10:00 level=INFO source=memory.go:326 msg="offload to cuda" layers.requested=-1 layers.model=61 layers.offload=10 layers.split="" memory.available="[23.4 GiB]" memory.gpu_overhead="0 B" memory.required.full="134.5 GiB" memory.required.partial="22.1 GiB" memory.required.kv="9.4 GiB" memory.required.allocations="[22.1 GiB]" memory.weights.total="132.5 GiB" memory.weights.repeating="132.1 GiB" memory.weights.nonrepeating="410.2 MiB" memory.graph.full="642.0 MiB" memory.graph.partial="891.5 MiB" Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.911+10:00 level=INFO source=server.go:388 msg="starting llama server" cmd="/tmp/ollama1470999858/runners/cuda_v12/ollama_llama_server --model /mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 10 --no-mmap --parallel 1 --port 43133" Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.911+10:00 level=INFO source=sched.go:449 msg="loaded runners" count=1 Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.911+10:00 level=INFO source=server.go:587 msg="waiting for llama runner to start responding" Oct 02 09:22:03 xlr2 ollama[2576620]: time=2024-10-02T09:22:03.912+10:00 level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server error" Oct 02 09:22:03 xlr2 ollama[3244519]: INFO [main] build info | build=10 commit="9225b05" tid="139874777862144" timestamp=1727824923 Oct 02 09:22:03 xlr2 ollama[3244519]: INFO [main] system info | n_threads=8 n_threads_batch=8 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="139874777862144" timestamp=1727824923 total_threads=16 Oct 02 09:22:03 xlr2 ollama[3244519]: INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="15" port="43133" tid="139874777862144" timestamp=1727824923 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: loaded meta data with 46 key-value pairs and 959 tensors from /mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c (version GGUF V3 (latest)) Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 0: general.architecture str = deepseek2 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 1: general.type str = model Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 2: general.name str = DeepSeek V2.5 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 3: general.version str = V2.5 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 4: general.basename str = DeepSeek Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 5: general.size_label str = 160x14B Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 6: general.license str = other Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 7: general.license.name str = deepseek Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 8: general.license.link str = https://github.com/deepseek-ai/DeepSe... Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 9: deepseek2.block_count u32 = 60 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 10: deepseek2.context_length u32 = 163840 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 11: deepseek2.embedding_length u32 = 5120 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 12: deepseek2.feed_forward_length u32 = 12288 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 13: deepseek2.attention.head_count u32 = 128 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 14: deepseek2.attention.head_count_kv u32 = 128 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 15: deepseek2.rope.freq_base f32 = 10000.000000 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 16: deepseek2.attention.layer_norm_rms_epsilon f32 = 0.000001 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 17: deepseek2.expert_used_count u32 = 6 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 18: general.file_type u32 = 2 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 19: deepseek2.leading_dense_block_count u32 = 1 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 20: deepseek2.vocab_size u32 = 102400 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 21: deepseek2.attention.q_lora_rank u32 = 1536 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 22: deepseek2.attention.kv_lora_rank u32 = 512 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 23: deepseek2.attention.key_length u32 = 192 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 24: deepseek2.attention.value_length u32 = 128 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 25: deepseek2.expert_feed_forward_length u32 = 1536 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 26: deepseek2.expert_count u32 = 160 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 27: deepseek2.expert_shared_count u32 = 2 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 28: deepseek2.expert_weights_scale f32 = 16.000000 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 29: deepseek2.rope.dimension_count u32 = 64 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 30: deepseek2.rope.scaling.type str = yarn Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 31: deepseek2.rope.scaling.factor f32 = 40.000000 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 32: deepseek2.rope.scaling.original_context_length u32 = 4096 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 33: deepseek2.rope.scaling.yarn_log_multiplier f32 = 0.100000 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 34: tokenizer.ggml.model str = gpt2 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 35: tokenizer.ggml.pre str = deepseek-llm Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 36: tokenizer.ggml.tokens arr[str,102400] = ["!", "\"", "#", "$", "%", "&", "'", ... Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 37: tokenizer.ggml.token_type arr[i32,102400] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 38: tokenizer.ggml.merges arr[str,99757] = ["G G", "G t", "G a", "i n", "h e... Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 39: tokenizer.ggml.bos_token_id u32 = 100000 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 40: tokenizer.ggml.eos_token_id u32 = 100001 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 41: tokenizer.ggml.padding_token_id u32 = 100001 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 42: tokenizer.ggml.add_bos_token bool = true Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 43: tokenizer.ggml.add_eos_token bool = false Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 44: tokenizer.chat_template str = {% if not add_generation_prompt is de... Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - kv 45: general.quantization_version u32 = 2 Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - type f32: 300 tensors Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - type q4_0: 658 tensors Oct 02 09:22:03 xlr2 ollama[2576620]: llama_model_loader: - type q6_K: 1 tensors Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_vocab: special tokens cache size = 18 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_vocab: token to piece cache size = 0.6411 MB Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: format = GGUF V3 (latest) Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: arch = deepseek2 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: vocab type = BPE Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_vocab = 102400 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_merges = 99757 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: vocab_only = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_ctx_train = 163840 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_embd = 5120 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_layer = 60 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_head = 128 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_head_kv = 128 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_rot = 64 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_swa = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_embd_head_k = 192 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_embd_head_v = 128 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_gqa = 1 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_embd_k_gqa = 24576 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_embd_v_gqa = 16384 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: f_norm_eps = 0.0e+00 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: f_norm_rms_eps = 1.0e-06 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: f_clamp_kqv = 0.0e+00 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: f_max_alibi_bias = 0.0e+00 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: f_logit_scale = 0.0e+00 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_ff = 12288 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_expert = 160 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_expert_used = 6 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: causal attn = 1 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: pooling type = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: rope type = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: rope scaling = yarn Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: freq_base_train = 10000.0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: freq_scale_train = 0.025 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_ctx_orig_yarn = 4096 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: rope_finetuned = unknown Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: ssm_d_conv = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: ssm_d_inner = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: ssm_d_state = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: ssm_dt_rank = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: ssm_dt_b_c_rms = 0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: model type = 236B Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: model ftype = Q4_0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: model params = 235.74 B Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: model size = 123.78 GiB (4.51 BPW) Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: general.name = DeepSeek V2.5 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: BOS token = 100000 '<|begin?of?sentence|>' Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: EOS token = 100001 '<|end?of?sentence|>' Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: PAD token = 100001 '<|end?of?sentence|>' Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: LF token = 126 'Ä' Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: max token length = 256 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_layer_dense_lead = 1 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_lora_q = 1536 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_lora_kv = 512 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_ff_exp = 1536 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: n_expert_shared = 2 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: expert_weights_scale = 16.0 Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_print_meta: rope_yarn_log_mul = 0.1000 Oct 02 09:22:04 xlr2 ollama[2576620]: time=2024-10-02T09:22:04.163+10:00 level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server loading model" Oct 02 09:22:04 xlr2 ollama[2576620]: ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no Oct 02 09:22:04 xlr2 ollama[2576620]: ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no Oct 02 09:22:04 xlr2 ollama[2576620]: ggml_cuda_init: found 1 CUDA devices: Oct 02 09:22:04 xlr2 ollama[2576620]: Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes Oct 02 09:22:04 xlr2 ollama[2576620]: llm_load_tensors: ggml ctx size = 0.80 MiB Oct 02 09:22:05 xlr2 ollama[2576620]: time=2024-10-02T09:22:05.618+10:00 level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server not responding" Oct 02 09:22:33 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:22:37 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:22:47 xlr2 ollama[2576620]: time=2024-10-02T09:22:47.399+10:00 level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server loading model" Oct 02 09:22:51 xlr2 ollama[2576620]: llm_load_tensors: offloading 10 repeating layers to GPU Oct 02 09:22:51 xlr2 ollama[2576620]: llm_load_tensors: offloaded 10/61 layers to GPU Oct 02 09:22:51 xlr2 ollama[2576620]: llm_load_tensors: CUDA_Host buffer size = 105416.00 MiB Oct 02 09:22:51 xlr2 ollama[2576620]: llm_load_tensors: CUDA0 buffer size = 21335.35 MiB Oct 02 09:23:10 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:23:17 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:23:44 xlr2 ollama[2576620]: llama_new_context_with_model: n_ctx = 2048 Oct 02 09:23:44 xlr2 ollama[2576620]: llama_new_context_with_model: n_batch = 512 Oct 02 09:23:44 xlr2 ollama[2576620]: llama_new_context_with_model: n_ubatch = 512 Oct 02 09:23:44 xlr2 ollama[2576620]: llama_new_context_with_model: flash_attn = 0 Oct 02 09:23:44 xlr2 ollama[2576620]: llama_new_context_with_model: freq_base = 10000.0 Oct 02 09:23:44 xlr2 ollama[2576620]: llama_new_context_with_model: freq_scale = 0.025 Oct 02 09:23:48 xlr2 ollama[2576620]: llama_kv_cache_init: CUDA_Host KV buffer size = 8000.00 MiB Oct 02 09:23:48 xlr2 ollama[2576620]: llama_kv_cache_init: CUDA0 KV buffer size = 1600.00 MiB Oct 02 09:23:48 xlr2 ollama[2576620]: llama_new_context_with_model: KV self size = 9600.00 MiB, K (f16): 5760.00 MiB, V (f16): 3840.00 MiB Oct 02 09:23:48 xlr2 ollama[2576620]: llama_new_context_with_model: CUDA_Host output buffer size = 0.41 MiB Oct 02 09:23:48 xlr2 ollama[2576620]: ggml_backend_cuda_buffer_type_alloc_buffer: allocating 842.00 MiB on device 0: cudaMalloc failed: out of memory Oct 02 09:23:48 xlr2 ollama[2576620]: ggml_gallocr_reserve_n: failed to allocate CUDA0 buffer of size 882903040 Oct 02 09:23:48 xlr2 ollama[2576620]: llama_new_context_with_model: failed to allocate compute buffers Oct 02 09:23:50 xlr2 ollama[2576620]: llama_init_from_gpt_params: error: failed to create context with model '/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c' Oct 02 09:23:52 xlr2 kernel: clocksource: Long readout interval, skipping watchdog check: cs_nsec: 1586308203 wd_nsec: 1586308525 Oct 02 09:24:05 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:24:09 xlr2 systemd-networkd[3394145]: enp7s0: DHCPv6 lease lost Oct 02 09:24:10 xlr2 ollama[3244519]: ERROR [load_model] unable to load model | model="/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c" tid="139874777862144" timestamp=1727825050 Oct 02 09:24:10 xlr2 ollama[2576620]: terminate called without an active exception Oct 02 09:24:10 xlr2 ollama[2576620]: time=2024-10-02T09:24:10.958+10:00 level=INFO source=server.go:621 msg="waiting for server to become available" status="llm server error" Oct 02 09:24:11 xlr2 ollama[2576620]: time=2024-10-02T09:24:11.213+10:00 level=ERROR source=sched.go:455 msg="error loading llama server" error="llama runner process has terminated: error:failed to create context with model '/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c'" Oct 02 09:24:11 xlr2 ollama[2576620]: [GIN] 2024/10/02 - 09:24:11 | 500 | 2m8s | 127.0.0.1 | POST "/api/generate" Oct 02 09:24:16 xlr2 ollama[2576620]: time=2024-10-02T09:24:16.336+10:00 level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=5.122496579 model=/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c Oct 02 09:24:16 xlr2 ollama[2576620]: time=2024-10-02T09:24:16.586+10:00 level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=5.372482202 model=/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c Oct 02 09:24:16 xlr2 ollama[2576620]: time=2024-10-02T09:24:16.837+10:00 level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=5.623037141 model=/mnt/ssd/ai/ollama/.ollama/models/blobs/sha256-799587243b19fdcc715a4aab927f5700d1b9508bd0b8b0db9dc2bd6fc622979c ``` System info: ``` $ uname -a Linux xlr2 5.15.0-112-generic #122-Ubuntu SMP Thu May 23 07:48:21 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux $ nproc 16 $ free -h total used free shared buff/cache available Mem: 125Gi 5.3Gi 113Gi 33Mi 6.5Gi 119Gi Swap: 8.0Gi 887Mi 7.1Gi $ nvidia-smi Wed Oct 2 10:04:06 2024 +---------------------------------------------------------------------------------------+ | NVIDIA-SMI 535.161.08 Driver Version: 535.161.08 CUDA Version: 12.2 | |-----------------------------------------+----------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+======================+======================| | 0 NVIDIA GeForce RTX 3090 On | 00000000:0C:00.0 Off | N/A | | 0% 33C P8 38W / 390W | 3MiB / 24576MiB | 0% Default | | | | N/A | +-----------------------------------------+----------------------+----------------------+ +---------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=======================================================================================| | No running processes found | +---------------------------------------------------------------------------------------+ $ ``` I tried with both `0.3.11` and the latest `0.3.12` versions of Ollama. Here is a snapshot of `htop` and `nvtop` outputs: <img width="820" alt="htop_nvtop" src="https://github.com/user-attachments/assets/e21266eb-1d3b-4ab6-9324-5604fc99efb4"> Here is the recording showing the entire process: https://github.com/user-attachments/assets/e306990a-a340-4a8f-b7de-b52d9b3de5e1 ### OS Linux ### GPU Nvidia ### CPU AMD ### Ollama version 0.3.11
{ "login": "LeonidShamis", "id": 1818114, "node_id": "MDQ6VXNlcjE4MTgxMTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/1818114?v=4", "gravatar_id": "", "url": "https://api.github.com/users/LeonidShamis", "html_url": "https://github.com/LeonidShamis", "followers_url": "https://api.github.com/users/LeonidShamis/followers", "following_url": "https://api.github.com/users/LeonidShamis/following{/other_user}", "gists_url": "https://api.github.com/users/LeonidShamis/gists{/gist_id}", "starred_url": "https://api.github.com/users/LeonidShamis/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/LeonidShamis/subscriptions", "organizations_url": "https://api.github.com/users/LeonidShamis/orgs", "repos_url": "https://api.github.com/users/LeonidShamis/repos", "events_url": "https://api.github.com/users/LeonidShamis/events{/privacy}", "received_events_url": "https://api.github.com/users/LeonidShamis/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7072/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7072/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8634
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8634/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8634/comments
https://api.github.com/repos/ollama/ollama/issues/8634/events
https://github.com/ollama/ollama/issues/8634
2,815,768,303
I_kwDOJ0Z1Ps6n1Tbv
8,634
Ollama is not installing on Termux
{ "login": "imvickykumar999", "id": 50515418, "node_id": "MDQ6VXNlcjUwNTE1NDE4", "avatar_url": "https://avatars.githubusercontent.com/u/50515418?v=4", "gravatar_id": "", "url": "https://api.github.com/users/imvickykumar999", "html_url": "https://github.com/imvickykumar999", "followers_url": "https://api.github.com/users/imvickykumar999/followers", "following_url": "https://api.github.com/users/imvickykumar999/following{/other_user}", "gists_url": "https://api.github.com/users/imvickykumar999/gists{/gist_id}", "starred_url": "https://api.github.com/users/imvickykumar999/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/imvickykumar999/subscriptions", "organizations_url": "https://api.github.com/users/imvickykumar999/orgs", "repos_url": "https://api.github.com/users/imvickykumar999/repos", "events_url": "https://api.github.com/users/imvickykumar999/events{/privacy}", "received_events_url": "https://api.github.com/users/imvickykumar999/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
3
2025-01-28T14:01:22
2025-01-30T07:10:12
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
~ $ curl -fsSL https://ollama.com/install.sh | sh ``` >>> Installing ollama to /usr No superuser binary detected. Are you rooted? ```
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8634/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8634/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6190
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6190/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6190/comments
https://api.github.com/repos/ollama/ollama/issues/6190/events
https://github.com/ollama/ollama/pull/6190
2,449,645,987
PR_kwDOJ0Z1Ps53gV83
6,190
fix concurrency test
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-08-05T23:36:34
2024-08-05T23:45:52
2024-08-05T23:45:50
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6190", "html_url": "https://github.com/ollama/ollama/pull/6190", "diff_url": "https://github.com/ollama/ollama/pull/6190.diff", "patch_url": "https://github.com/ollama/ollama/pull/6190.patch", "merged_at": "2024-08-05T23:45:50" }
errors were hidden by `integration` build tag
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6190/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6190/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8350
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8350/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8350/comments
https://api.github.com/repos/ollama/ollama/issues/8350/events
https://github.com/ollama/ollama/pull/8350
2,776,137,506
PR_kwDOJ0Z1Ps6HHapv
8,350
readme: add phi4 model
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2025-01-08T19:17:03
2025-01-08T19:21:41
2025-01-08T19:21:39
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8350", "html_url": "https://github.com/ollama/ollama/pull/8350", "diff_url": "https://github.com/ollama/ollama/pull/8350.diff", "patch_url": "https://github.com/ollama/ollama/pull/8350.patch", "merged_at": "2025-01-08T19:21:39" }
readme: add phi4 model
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8350/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8350/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8689
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8689/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8689/comments
https://api.github.com/repos/ollama/ollama/issues/8689/events
https://github.com/ollama/ollama/issues/8689
2,820,234,513
I_kwDOJ0Z1Ps6oGV0R
8,689
Error LLama runner process has terminated: %!w(<nil>)
{ "login": "Saatvik-droid", "id": 55750489, "node_id": "MDQ6VXNlcjU1NzUwNDg5", "avatar_url": "https://avatars.githubusercontent.com/u/55750489?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Saatvik-droid", "html_url": "https://github.com/Saatvik-droid", "followers_url": "https://api.github.com/users/Saatvik-droid/followers", "following_url": "https://api.github.com/users/Saatvik-droid/following{/other_user}", "gists_url": "https://api.github.com/users/Saatvik-droid/gists{/gist_id}", "starred_url": "https://api.github.com/users/Saatvik-droid/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Saatvik-droid/subscriptions", "organizations_url": "https://api.github.com/users/Saatvik-droid/orgs", "repos_url": "https://api.github.com/users/Saatvik-droid/repos", "events_url": "https://api.github.com/users/Saatvik-droid/events{/privacy}", "received_events_url": "https://api.github.com/users/Saatvik-droid/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
1
2025-01-30T08:49:09
2025-01-30T08:57:59
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Sometimes when infering from ollama using the python module I get this error. After retrying a couple of times it works and looks random to me. ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.7
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8689/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8689/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/3376
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3376/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3376/comments
https://api.github.com/repos/ollama/ollama/issues/3376/events
https://github.com/ollama/ollama/pull/3376
2,211,769,650
PR_kwDOJ0Z1Ps5q-g51
3,376
only generate on changes to llm subdirectory
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-27T19:45:40
2024-03-27T21:12:54
2024-03-27T21:12:53
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/3376", "html_url": "https://github.com/ollama/ollama/pull/3376", "diff_url": "https://github.com/ollama/ollama/pull/3376.diff", "patch_url": "https://github.com/ollama/ollama/pull/3376.patch", "merged_at": "2024-03-27T21:12:53" }
follow up to #3375 to also skip generate (linux, macos, windows) if there's no changes to the llm subdirectory
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3376/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3376/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7374
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7374/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7374/comments
https://api.github.com/repos/ollama/ollama/issues/7374/events
https://github.com/ollama/ollama/issues/7374
2,616,050,717
I_kwDOJ0Z1Ps6b7cQd
7,374
Reinstate OLLAMA_RUNNERS_DIR
{ "login": "StarPet", "id": 85790781, "node_id": "MDQ6VXNlcjg1NzkwNzgx", "avatar_url": "https://avatars.githubusercontent.com/u/85790781?v=4", "gravatar_id": "", "url": "https://api.github.com/users/StarPet", "html_url": "https://github.com/StarPet", "followers_url": "https://api.github.com/users/StarPet/followers", "following_url": "https://api.github.com/users/StarPet/following{/other_user}", "gists_url": "https://api.github.com/users/StarPet/gists{/gist_id}", "starred_url": "https://api.github.com/users/StarPet/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/StarPet/subscriptions", "organizations_url": "https://api.github.com/users/StarPet/orgs", "repos_url": "https://api.github.com/users/StarPet/repos", "events_url": "https://api.github.com/users/StarPet/events{/privacy}", "received_events_url": "https://api.github.com/users/StarPet/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
5
2024-10-26T18:28:24
2024-11-06T15:38:33
2024-11-06T15:38:33
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
It appears that the OLLAMA_RUNNERS_DIR was removed from the code - at least I couldn't find it in github's search function. Currently (0.3.14) it is using /tmp/ollama<number>/runners again, as before the introduction of the OLLAMA_RUNNERS_DIR (or when not set). IMHO, using /tmp for executables is not a good idea. I'd prefer to have some control over where the executables are extracted to. So, please, bring back the OLLAMA_RUNNERS_DIR.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7374/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7374/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3482
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3482/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3482/comments
https://api.github.com/repos/ollama/ollama/issues/3482/events
https://github.com/ollama/ollama/issues/3482
2,224,411,948
I_kwDOJ0Z1Ps6EldUs
3,482
Please add Qwen-VL!
{ "login": "tikeoewoew", "id": 12619882, "node_id": "MDQ6VXNlcjEyNjE5ODgy", "avatar_url": "https://avatars.githubusercontent.com/u/12619882?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tikeoewoew", "html_url": "https://github.com/tikeoewoew", "followers_url": "https://api.github.com/users/tikeoewoew/followers", "following_url": "https://api.github.com/users/tikeoewoew/following{/other_user}", "gists_url": "https://api.github.com/users/tikeoewoew/gists{/gist_id}", "starred_url": "https://api.github.com/users/tikeoewoew/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/tikeoewoew/subscriptions", "organizations_url": "https://api.github.com/users/tikeoewoew/orgs", "repos_url": "https://api.github.com/users/tikeoewoew/repos", "events_url": "https://api.github.com/users/tikeoewoew/events{/privacy}", "received_events_url": "https://api.github.com/users/tikeoewoew/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
2
2024-04-04T03:59:14
2024-07-29T01:25:43
2024-07-25T15:15:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What model would you like? This image recognition model is very popular in China, so please add it to ollama:https://huggingface.co/Qwen/Qwen-VL
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3482/reactions", "total_count": 10, "+1": 10, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3482/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/595
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/595/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/595/comments
https://api.github.com/repos/ollama/ollama/issues/595/events
https://github.com/ollama/ollama/pull/595
1,912,389,172
PR_kwDOJ0Z1Ps5bLAcR
595
ignore systemctl is-system-running exit code
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-25T22:47:54
2023-09-25T22:49:47
2023-09-25T22:49:47
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/595", "html_url": "https://github.com/ollama/ollama/pull/595", "diff_url": "https://github.com/ollama/ollama/pull/595.diff", "patch_url": "https://github.com/ollama/ollama/pull/595.patch", "merged_at": "2023-09-25T22:49:47" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/595/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/595/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8014
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8014/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8014/comments
https://api.github.com/repos/ollama/ollama/issues/8014/events
https://github.com/ollama/ollama/pull/8014
2,727,590,489
PR_kwDOJ0Z1Ps6EkWZ1
8,014
Avoid underflow when FreeMemory < overhead
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-12-09T16:15:23
2024-12-10T17:10:40
2024-12-10T17:10:40
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8014", "html_url": "https://github.com/ollama/ollama/pull/8014", "diff_url": "https://github.com/ollama/ollama/pull/8014.diff", "patch_url": "https://github.com/ollama/ollama/pull/8014.patch", "merged_at": "2024-12-10T17:10:40" }
Fixes: #8011
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8014/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8014/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1031
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1031/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1031/comments
https://api.github.com/repos/ollama/ollama/issues/1031/events
https://github.com/ollama/ollama/pull/1031
1,981,518,033
PR_kwDOJ0Z1Ps5e0LjA
1,031
Added logit_bias support
{ "login": "Vokturz", "id": 21696514, "node_id": "MDQ6VXNlcjIxNjk2NTE0", "avatar_url": "https://avatars.githubusercontent.com/u/21696514?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Vokturz", "html_url": "https://github.com/Vokturz", "followers_url": "https://api.github.com/users/Vokturz/followers", "following_url": "https://api.github.com/users/Vokturz/following{/other_user}", "gists_url": "https://api.github.com/users/Vokturz/gists{/gist_id}", "starred_url": "https://api.github.com/users/Vokturz/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Vokturz/subscriptions", "organizations_url": "https://api.github.com/users/Vokturz/orgs", "repos_url": "https://api.github.com/users/Vokturz/repos", "events_url": "https://api.github.com/users/Vokturz/events{/privacy}", "received_events_url": "https://api.github.com/users/Vokturz/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
2
2023-11-07T14:38:14
2024-03-16T16:23:18
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1031", "html_url": "https://github.com/ollama/ollama/pull/1031", "diff_url": "https://github.com/ollama/ollama/pull/1031.diff", "patch_url": "https://github.com/ollama/ollama/pull/1031.patch", "merged_at": null }
This PR brings the `logit_bias` functionality, already present in llama.cpp, which allows users to adjust the likelihood of token occurrences in generated text. For example, for the prompt `"Once upon a "` we have: 1. Without `logit_bias` ```bash curl -X POST http://localhost:11434/api/generate -d '{ "model": "llama2:7b", "prompt": "Once upon a ", "stream": false, "options": {"seed": 0, "num_predict" : 10} }' {"model":"llama2:7b", ... , "response":"time, in a far-off land, there", ...} ``` 2. With `logit_bias` removing "time" and "Time" words: ```bash curl -X POST http://localhost:11434/api/generate -d '{ "model": "llama2:7b", "prompt": "Once upon a ", "stream": false, "options": {"seed": 0, "num_predict" : 10, "logit_bias": [[2230,false],[931,false],[2481,false],[5974,false]]} }' {"model":"llama2:7b", ... , "response":"Once upon a...! What a great prompt!", ...} ``` 2. With `logit_bias` removing "time" and "Time" words, and increasing "night" likelihood : ```bash curl -X POST http://localhost:11434/api/generate -d '{ "model": "llama2:7b", "prompt": "Once upon a ", "stream": false, "options": {"seed": 0, "num_predict" : 10, "logit_bias": [[2230,false],[931,false],[2481,false],[5974,false],[11147, 15]]} }' {"model":"llama2:7b", ... , "response":"night, in a land far away, there lived"", ...} ```
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1031/reactions", "total_count": 9, "+1": 6, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 3, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1031/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/2628
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2628/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2628/comments
https://api.github.com/repos/ollama/ollama/issues/2628/events
https://github.com/ollama/ollama/issues/2628
2,146,310,477
I_kwDOJ0Z1Ps5_7hlN
2,628
libext_server.a(llava.cpp.o) { in archive is not an object | not an ELF file }
{ "login": "pavelsr", "id": 1158473, "node_id": "MDQ6VXNlcjExNTg0NzM=", "avatar_url": "https://avatars.githubusercontent.com/u/1158473?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pavelsr", "html_url": "https://github.com/pavelsr", "followers_url": "https://api.github.com/users/pavelsr/followers", "following_url": "https://api.github.com/users/pavelsr/following{/other_user}", "gists_url": "https://api.github.com/users/pavelsr/gists{/gist_id}", "starred_url": "https://api.github.com/users/pavelsr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pavelsr/subscriptions", "organizations_url": "https://api.github.com/users/pavelsr/orgs", "repos_url": "https://api.github.com/users/pavelsr/repos", "events_url": "https://api.github.com/users/pavelsr/events{/privacy}", "received_events_url": "https://api.github.com/users/pavelsr/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-02-21T09:57:34
2024-03-27T20:53:23
2024-03-27T20:53:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I'm trying to build ollama with AMD GPU support via command ``` ROCM_PATH=/opt/rocm CLBlast_DIR=/usr/lib/x86_64-linux-gnu/cmake/CLBlast go generate -tags rocm ./... ``` and in final stage of build I got an error mentioned in issue header: ![Screenshot from 2024-02-21 12-55-49](https://github.com/ollama/ollama/assets/1158473/93468605-6b88-4029-a890-793d135e6e48) I tried to replace `g++` with `clang++` at llm/generate/gen_common.sh as it was mentioned in [this comment](https://github.com/ollama/ollama/issues/2473#issuecomment-1942090475) but error in the end of build is still same. [Adding](https://github.com/ollama/ollama/issues/2473#issuecomment-1941640723) `-fcf-protection=none` also wasn't helped, just changed error message a little bit ``` ld.lld: error: ../llama.cpp/build/linux/x86_64/cpu/examples/server/libext_server.a(llava.cpp.o): not an ELF file clang++: error: linker command failed with exit code 1 (use -v to see invocation) ``` What do I need to fix for proper compiling of ext_server static library ? Seems like issue related to [this part of ollama sources](https://github.com/ollama/ollama/blob/e1f50377f499b1784a866a6463af364987443238/llm/generate/gen_common.sh#L89)
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2628/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2628/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1414
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1414/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1414/comments
https://api.github.com/repos/ollama/ollama/issues/1414/events
https://github.com/ollama/ollama/issues/1414
2,029,939,345
I_kwDOJ0Z1Ps54_mqR
1,414
Windows install runs into errors
{ "login": "csaben", "id": 76020733, "node_id": "MDQ6VXNlcjc2MDIwNzMz", "avatar_url": "https://avatars.githubusercontent.com/u/76020733?v=4", "gravatar_id": "", "url": "https://api.github.com/users/csaben", "html_url": "https://github.com/csaben", "followers_url": "https://api.github.com/users/csaben/followers", "following_url": "https://api.github.com/users/csaben/following{/other_user}", "gists_url": "https://api.github.com/users/csaben/gists{/gist_id}", "starred_url": "https://api.github.com/users/csaben/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/csaben/subscriptions", "organizations_url": "https://api.github.com/users/csaben/orgs", "repos_url": "https://api.github.com/users/csaben/repos", "events_url": "https://api.github.com/users/csaben/events{/privacy}", "received_events_url": "https://api.github.com/users/csaben/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
4
2023-12-07T05:33:03
2023-12-11T15:48:16
2023-12-11T15:48:16
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
after manual install instructions and ```go build``` I receive the following errors: ``` # github.com/jmorganca/ollama/readline readline\readline.go:199:12: undefined: syscall.Kill readline\readline.go:199:28: undefined: syscall.SIGSTOP ``` After messing with readline.go I just manage to get a .exe file that doesn't work @deadcoder0904 or anyone who compiled on windows and got a working executable, could you describe how you did it?
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1414/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1414/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/377
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/377/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/377/comments
https://api.github.com/repos/ollama/ollama/issues/377/events
https://github.com/ollama/ollama/pull/377
1,855,878,607
PR_kwDOJ0Z1Ps5YNGuk
377
Strip protocol from model path
{ "login": "rlbaker", "id": 967417, "node_id": "MDQ6VXNlcjk2NzQxNw==", "avatar_url": "https://avatars.githubusercontent.com/u/967417?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rlbaker", "html_url": "https://github.com/rlbaker", "followers_url": "https://api.github.com/users/rlbaker/followers", "following_url": "https://api.github.com/users/rlbaker/following{/other_user}", "gists_url": "https://api.github.com/users/rlbaker/gists{/gist_id}", "starred_url": "https://api.github.com/users/rlbaker/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rlbaker/subscriptions", "organizations_url": "https://api.github.com/users/rlbaker/orgs", "repos_url": "https://api.github.com/users/rlbaker/repos", "events_url": "https://api.github.com/users/rlbaker/events{/privacy}", "received_events_url": "https://api.github.com/users/rlbaker/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
5
2023-08-18T00:40:27
2023-08-22T04:56:57
2023-08-22T04:56:57
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/377", "html_url": "https://github.com/ollama/ollama/pull/377", "diff_url": "https://github.com/ollama/ollama/pull/377.diff", "patch_url": "https://github.com/ollama/ollama/pull/377.patch", "merged_at": "2023-08-22T04:56:57" }
Took a whack at fixing https://github.com/jmorganca/ollama/issues/371 and reorganized the switch logic slightly as well. Wasn't sure if it was better to strip all protocols or just `https://`, so if you'd like the latter I can switch it to just a `strings.TrimPrefix`. Happy to back out the updated switch code as well, just figured I'd do it while I was in there. Thanks for a great tool, loving it so far. Hope this is helpful.
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/377/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/377/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6312
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6312/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6312/comments
https://api.github.com/repos/ollama/ollama/issues/6312/events
https://github.com/ollama/ollama/issues/6312
2,459,666,429
I_kwDOJ0Z1Ps6Sm4f9
6,312
how to force ollama to use different cpu runners / how to compile windows avx512 runner?
{ "login": "AncientMystic", "id": 62780271, "node_id": "MDQ6VXNlcjYyNzgwMjcx", "avatar_url": "https://avatars.githubusercontent.com/u/62780271?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AncientMystic", "html_url": "https://github.com/AncientMystic", "followers_url": "https://api.github.com/users/AncientMystic/followers", "following_url": "https://api.github.com/users/AncientMystic/following{/other_user}", "gists_url": "https://api.github.com/users/AncientMystic/gists{/gist_id}", "starred_url": "https://api.github.com/users/AncientMystic/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AncientMystic/subscriptions", "organizations_url": "https://api.github.com/users/AncientMystic/orgs", "repos_url": "https://api.github.com/users/AncientMystic/repos", "events_url": "https://api.github.com/users/AncientMystic/events{/privacy}", "received_events_url": "https://api.github.com/users/AncientMystic/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 5808482718, "node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng", "url": "https://api.github.com/repos/ollama/ollama/labels/performance", "name": "performance", "color": "A5B5C6", "default": false, "description": "" }, { "id": 6430601766, "node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg", "url": "https://api.github.com/repos/ollama/ollama/labels/nvidia", "name": "nvidia", "color": "8CDB00", "default": false, "description": "Issues relating to Nvidia GPUs and CUDA" }, { "id": 7700262114, "node_id": "LA_kwDOJ0Z1Ps8AAAAByvis4g", "url": "https://api.github.com/repos/ollama/ollama/labels/build", "name": "build", "color": "006b75", "default": false, "description": "Issues relating to building ollama from source" } ]
open
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
16
2024-08-11T16:31:17
2024-11-04T19:16:38
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? according to logs ollama seems to only be using AVX not AVX2, how would i fix this and force avx2 or higher? also wondering how i compile the avx512 runner for windows? i have compiled other runners and the cuda runner fine but it seems regardless of what i try to set it just generates cpu, avx, avx2 and stops running `$env:OLLAMA_CUSTOM_CPU_DEFS="-DGGML_AVX=on -DGGML_AVX2=on -DGGML_AVX512=on"` or any other combo of this command seems to have no effect on generating a different runner aside from defaults, is there a file i need to edit to change what is compiled? i would also like to skip cpu, avx, avx2 if possible as it takes a fair amount of time and those are already compiled cpu is i7-7820x with support for avx2 & avx512 ### OS Windows ### GPU Nvidia, Intel ### CPU Intel ### Ollama version 0.3.4
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6312/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6312/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/6545
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6545/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6545/comments
https://api.github.com/repos/ollama/ollama/issues/6545/events
https://github.com/ollama/ollama/pull/6545
2,492,937,842
PR_kwDOJ0Z1Ps55wQ1a
6,545
add llama3.1 chat template
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-08-28T20:30:23
2024-08-28T21:03:22
2024-08-28T21:03:20
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6545", "html_url": "https://github.com/ollama/ollama/pull/6545", "diff_url": "https://github.com/ollama/ollama/pull/6545.diff", "patch_url": "https://github.com/ollama/ollama/pull/6545.patch", "merged_at": "2024-08-28T21:03:20" }
This change adds the llama3.1 chat template do that it will be autodetected in `ollama create`.
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6545/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6545/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7665
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7665/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7665/comments
https://api.github.com/repos/ollama/ollama/issues/7665/events
https://github.com/ollama/ollama/issues/7665
2,658,023,476
I_kwDOJ0Z1Ps6ebjg0
7,665
Dealing with passing huge attachments to models?
{ "login": "robotom", "id": 45123215, "node_id": "MDQ6VXNlcjQ1MTIzMjE1", "avatar_url": "https://avatars.githubusercontent.com/u/45123215?v=4", "gravatar_id": "", "url": "https://api.github.com/users/robotom", "html_url": "https://github.com/robotom", "followers_url": "https://api.github.com/users/robotom/followers", "following_url": "https://api.github.com/users/robotom/following{/other_user}", "gists_url": "https://api.github.com/users/robotom/gists{/gist_id}", "starred_url": "https://api.github.com/users/robotom/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/robotom/subscriptions", "organizations_url": "https://api.github.com/users/robotom/orgs", "repos_url": "https://api.github.com/users/robotom/repos", "events_url": "https://api.github.com/users/robotom/events{/privacy}", "received_events_url": "https://api.github.com/users/robotom/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
1
2024-11-14T08:36:09
2024-11-17T12:24:08
2024-11-17T12:24:08
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi all, I'm wondering what the best way is to pass huge amounts of "attachment" data to any of the models? A typical doc might be > 2 million words (~16 million characters). I know this amount of data can't be fed to any under 70B as prompts due to context windows. If I go with llama 3.1 405B on a 4 x H100 NVL or 8 x H100 NVL, would that be the smart way to go? Or is there a better approach to this that I'm missing (like training my own model or something). Ideally I'd want to be able to just drop one of these docs into the model at any time as these files come new to me, frequently. Thanks!
{ "login": "rick-github", "id": 14946854, "node_id": "MDQ6VXNlcjE0OTQ2ODU0", "avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rick-github", "html_url": "https://github.com/rick-github", "followers_url": "https://api.github.com/users/rick-github/followers", "following_url": "https://api.github.com/users/rick-github/following{/other_user}", "gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}", "starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rick-github/subscriptions", "organizations_url": "https://api.github.com/users/rick-github/orgs", "repos_url": "https://api.github.com/users/rick-github/repos", "events_url": "https://api.github.com/users/rick-github/events{/privacy}", "received_events_url": "https://api.github.com/users/rick-github/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7665/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7665/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5585
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5585/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5585/comments
https://api.github.com/repos/ollama/ollama/issues/5585/events
https://github.com/ollama/ollama/pull/5585
2,399,405,473
PR_kwDOJ0Z1Ps505WvW
5,585
Create SECURITY.md
{ "login": "thejefflarson", "id": 55365, "node_id": "MDQ6VXNlcjU1MzY1", "avatar_url": "https://avatars.githubusercontent.com/u/55365?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thejefflarson", "html_url": "https://github.com/thejefflarson", "followers_url": "https://api.github.com/users/thejefflarson/followers", "following_url": "https://api.github.com/users/thejefflarson/following{/other_user}", "gists_url": "https://api.github.com/users/thejefflarson/gists{/gist_id}", "starred_url": "https://api.github.com/users/thejefflarson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thejefflarson/subscriptions", "organizations_url": "https://api.github.com/users/thejefflarson/orgs", "repos_url": "https://api.github.com/users/thejefflarson/repos", "events_url": "https://api.github.com/users/thejefflarson/events{/privacy}", "received_events_url": "https://api.github.com/users/thejefflarson/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-07-09T23:25:27
2024-08-02T22:19:44
2024-08-02T16:53:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5585", "html_url": "https://github.com/ollama/ollama/pull/5585", "diff_url": "https://github.com/ollama/ollama/pull/5585.diff", "patch_url": "https://github.com/ollama/ollama/pull/5585.patch", "merged_at": null }
I know that Ollama is under active development, but it would be great to have a security policy in place for folks to report bugs. Also, would Ollama be open to enabling code scanning? https://docs.github.com/en/code-security/code-scanning/enabling-code-scanning/configuring-default-setup-for-code-scanning
{ "login": "thejefflarson", "id": 55365, "node_id": "MDQ6VXNlcjU1MzY1", "avatar_url": "https://avatars.githubusercontent.com/u/55365?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thejefflarson", "html_url": "https://github.com/thejefflarson", "followers_url": "https://api.github.com/users/thejefflarson/followers", "following_url": "https://api.github.com/users/thejefflarson/following{/other_user}", "gists_url": "https://api.github.com/users/thejefflarson/gists{/gist_id}", "starred_url": "https://api.github.com/users/thejefflarson/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/thejefflarson/subscriptions", "organizations_url": "https://api.github.com/users/thejefflarson/orgs", "repos_url": "https://api.github.com/users/thejefflarson/repos", "events_url": "https://api.github.com/users/thejefflarson/events{/privacy}", "received_events_url": "https://api.github.com/users/thejefflarson/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5585/reactions", "total_count": 2, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5585/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6540
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6540/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6540/comments
https://api.github.com/repos/ollama/ollama/issues/6540/events
https://github.com/ollama/ollama/issues/6540
2,491,433,636
I_kwDOJ0Z1Ps6UgEKk
6,540
actively retrieves the content returned from the web page
{ "login": "Nurburgring-Zhang", "id": 171787109, "node_id": "U_kgDOCj1DZQ", "avatar_url": "https://avatars.githubusercontent.com/u/171787109?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nurburgring-Zhang", "html_url": "https://github.com/Nurburgring-Zhang", "followers_url": "https://api.github.com/users/Nurburgring-Zhang/followers", "following_url": "https://api.github.com/users/Nurburgring-Zhang/following{/other_user}", "gists_url": "https://api.github.com/users/Nurburgring-Zhang/gists{/gist_id}", "starred_url": "https://api.github.com/users/Nurburgring-Zhang/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nurburgring-Zhang/subscriptions", "organizations_url": "https://api.github.com/users/Nurburgring-Zhang/orgs", "repos_url": "https://api.github.com/users/Nurburgring-Zhang/repos", "events_url": "https://api.github.com/users/Nurburgring-Zhang/events{/privacy}", "received_events_url": "https://api.github.com/users/Nurburgring-Zhang/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
1
2024-08-28T08:31:10
2024-08-28T16:35:48
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
expected that ollama can automatically identify the model, and then when the problem exceeds the capacity of the model, ollama actively retrieves the content returned from the web page to the model, and the model analyzes the content returned and finally gives the answer.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6540/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6540/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/4763
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4763/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4763/comments
https://api.github.com/repos/ollama/ollama/issues/4763/events
https://github.com/ollama/ollama/issues/4763
2,329,008,117
I_kwDOJ0Z1Ps6K0df1
4,763
I created Ollama - Open WebUI Script - Give it a try!
{ "login": "Special-Niewbie", "id": 64843123, "node_id": "MDQ6VXNlcjY0ODQzMTIz", "avatar_url": "https://avatars.githubusercontent.com/u/64843123?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Special-Niewbie", "html_url": "https://github.com/Special-Niewbie", "followers_url": "https://api.github.com/users/Special-Niewbie/followers", "following_url": "https://api.github.com/users/Special-Niewbie/following{/other_user}", "gists_url": "https://api.github.com/users/Special-Niewbie/gists{/gist_id}", "starred_url": "https://api.github.com/users/Special-Niewbie/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Special-Niewbie/subscriptions", "organizations_url": "https://api.github.com/users/Special-Niewbie/orgs", "repos_url": "https://api.github.com/users/Special-Niewbie/repos", "events_url": "https://api.github.com/users/Special-Niewbie/events{/privacy}", "received_events_url": "https://api.github.com/users/Special-Niewbie/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
1
2024-06-01T08:31:52
2024-09-14T17:13:43
2024-09-14T17:13:42
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I created "Ollama-Open-WebUI-Script" for those who, like me, prefer not to overload their PC at startup with too many resources. Instead of manually starting Docker, then Ollama, and finally Open WebUI, this script simplifies the entire process, giving you direct and easy access to these two fantastic projects. Give it a try here https://github.com/Special-Niewbie/Ollama-Open-WebUI-Script, and enjoy a smoother experience with Ollama and Open WebUI!
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4763/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/4763/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6584
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6584/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6584/comments
https://api.github.com/repos/ollama/ollama/issues/6584/events
https://github.com/ollama/ollama/pull/6584
2,499,328,604
PR_kwDOJ0Z1Ps56E8ab
6,584
Add serve step to quickstart
{ "login": "anitagraser", "id": 590385, "node_id": "MDQ6VXNlcjU5MDM4NQ==", "avatar_url": "https://avatars.githubusercontent.com/u/590385?v=4", "gravatar_id": "", "url": "https://api.github.com/users/anitagraser", "html_url": "https://github.com/anitagraser", "followers_url": "https://api.github.com/users/anitagraser/followers", "following_url": "https://api.github.com/users/anitagraser/following{/other_user}", "gists_url": "https://api.github.com/users/anitagraser/gists{/gist_id}", "starred_url": "https://api.github.com/users/anitagraser/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/anitagraser/subscriptions", "organizations_url": "https://api.github.com/users/anitagraser/orgs", "repos_url": "https://api.github.com/users/anitagraser/repos", "events_url": "https://api.github.com/users/anitagraser/events{/privacy}", "received_events_url": "https://api.github.com/users/anitagraser/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-09-01T09:11:18
2024-09-02T19:42:04
2024-09-02T19:32:39
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/6584", "html_url": "https://github.com/ollama/ollama/pull/6584", "diff_url": "https://github.com/ollama/ollama/pull/6584.diff", "patch_url": "https://github.com/ollama/ollama/pull/6584.patch", "merged_at": null }
This missing step trips up beginners, as shown in https://github.com/ollama/ollama/issues/2727#issuecomment-1969331044
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6584/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6584/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5735
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5735/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5735/comments
https://api.github.com/repos/ollama/ollama/issues/5735/events
https://github.com/ollama/ollama/issues/5735
2,412,328,010
I_kwDOJ0Z1Ps6PyTRK
5,735
How to Set Up RAG / LLamaIndex with Windows Preview?
{ "login": "elikakohen", "id": 11563283, "node_id": "MDQ6VXNlcjExNTYzMjgz", "avatar_url": "https://avatars.githubusercontent.com/u/11563283?v=4", "gravatar_id": "", "url": "https://api.github.com/users/elikakohen", "html_url": "https://github.com/elikakohen", "followers_url": "https://api.github.com/users/elikakohen/followers", "following_url": "https://api.github.com/users/elikakohen/following{/other_user}", "gists_url": "https://api.github.com/users/elikakohen/gists{/gist_id}", "starred_url": "https://api.github.com/users/elikakohen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/elikakohen/subscriptions", "organizations_url": "https://api.github.com/users/elikakohen/orgs", "repos_url": "https://api.github.com/users/elikakohen/repos", "events_url": "https://api.github.com/users/elikakohen/events{/privacy}", "received_events_url": "https://api.github.com/users/elikakohen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
0
2024-07-17T00:59:03
2024-07-17T01:00:45
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I was wondering how, or if there is a way, to set up RAG with the Windows version?
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5735/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5735/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/2627
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2627/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2627/comments
https://api.github.com/repos/ollama/ollama/issues/2627/events
https://github.com/ollama/ollama/issues/2627
2,146,269,379
I_kwDOJ0Z1Ps5_7XjD
2,627
Error: listen tcp 127.0.0.1:11434: bind:
{ "login": "szymonk92", "id": 4785319, "node_id": "MDQ6VXNlcjQ3ODUzMTk=", "avatar_url": "https://avatars.githubusercontent.com/u/4785319?v=4", "gravatar_id": "", "url": "https://api.github.com/users/szymonk92", "html_url": "https://github.com/szymonk92", "followers_url": "https://api.github.com/users/szymonk92/followers", "following_url": "https://api.github.com/users/szymonk92/following{/other_user}", "gists_url": "https://api.github.com/users/szymonk92/gists{/gist_id}", "starred_url": "https://api.github.com/users/szymonk92/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/szymonk92/subscriptions", "organizations_url": "https://api.github.com/users/szymonk92/orgs", "repos_url": "https://api.github.com/users/szymonk92/repos", "events_url": "https://api.github.com/users/szymonk92/events{/privacy}", "received_events_url": "https://api.github.com/users/szymonk92/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 5860134234, "node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg", "url": "https://api.github.com/repos/ollama/ollama/labels/windows", "name": "windows", "color": "0052CC", "default": false, "description": "" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
13
2024-02-21T09:41:05
2024-09-29T01:41:48
2024-03-27T20:54:57
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Windows 10, I cannot start Ollama, ``` $ ollama serve Error: listen tcp 127.0.0.1:11434: bind: An attempt was made to access a socket in a way forbidden by its access permissions. ``` from **app.log** ``` time=2024-02-21T10:04:42.504+01:00 level=WARN source=server.go:109 msg="server crash 332 - exit code 1 - respawning" time=2024-02-21T10:04:43.018+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" time=2024-02-21T10:07:58.072+01:00 level=DEBUG source=logging_windows.go:12 msg="viewing logs with start C:\\Users\\user\\AppData\\Local\\Ollama" time=2024-02-21T10:10:15.024+01:00 level=WARN source=server.go:109 msg="server crash 333 - exit code 1 - respawning" time=2024-02-21T10:10:15.526+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" time=2024-02-21T10:11:20.094+01:00 level=DEBUG source=eventloop.go:151 msg="unmanaged app message, lParm: 0x204" time=2024-02-21T10:11:20.792+01:00 level=DEBUG source=lifecycle.go:30 msg="QUIT called" time=2024-02-21T10:11:20.842+01:00 level=INFO source=lifecycle.go:78 msg="Waiting for ollama server to shutdown..." time=2024-02-21T10:15:48.537+01:00 level=DEBUG source=server.go:104 msg="server shutdown with exit code 1" time=2024-02-21T10:15:48.537+01:00 level=INFO source=lifecycle.go:82 msg="Ollama app exiting" time=2024-02-21T10:37:42.782+01:00 level=INFO source=logging.go:45 msg="ollama app started" time=2024-02-21T10:37:42.863+01:00 level=DEBUG source=lifecycle.go:26 msg="starting callback loop" time=2024-02-21T10:37:42.863+01:00 level=DEBUG source=store.go:61 msg="loaded existing store C:\\Users\\user\\AppData\\Local\\Ollama\\config.json - ID: 6fbe69d4-62a6-4ad8-9768-da087897ebe2" time=2024-02-21T10:37:42.864+01:00 level=DEBUG source=lifecycle.go:57 msg="Not first time, skipping first run notification" time=2024-02-21T10:37:42.865+01:00 level=DEBUG source=server.go:130 msg="heartbeat from server: Head \"http://127.0.0.1:11434/\": dial tcp 127.0.0.1:11434: connectex: No connection could be made because the target machine actively refused it." time=2024-02-21T10:37:42.865+01:00 level=INFO source=server.go:131 msg="unable to connect to server" time=2024-02-21T10:37:42.870+01:00 level=INFO source=server.go:87 msg="started ollama server with pid 40576" time=2024-02-21T10:37:42.870+01:00 level=INFO source=server.go:89 msg="ollama server logs C:\\Users\\user\\AppData\\Local\\Ollama\\server.log" time=2024-02-21T10:37:42.870+01:00 level=DEBUG source=eventloop.go:24 msg="starting event handling loop" time=2024-02-21T10:37:42.941+01:00 level=WARN source=server.go:109 msg="server crash 1 - exit code 1 - respawning" time=2024-02-21T10:37:43.445+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" time=2024-02-21T10:37:44.461+01:00 level=WARN source=server.go:109 msg="server crash 2 - exit code 1 - respawning" time=2024-02-21T10:37:44.636+01:00 level=WARN source=tray.go:365 msg="failed to bring menu to foreground: The operation completed successfully." time=2024-02-21T10:37:44.971+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" time=2024-02-21T10:37:45.525+01:00 level=DEBUG source=logging_windows.go:12 msg="viewing logs with start C:\\Users\\user\\AppData\\Local\\Ollama" time=2024-02-21T10:37:45.882+01:00 level=DEBUG source=updater.go:87 msg="checking for available update at https://ollama.com/api/update?arch=amd64&nonce=RENbdDvJrUl9L0FRExH2xw&os=windows&ts=1708508265&version=0.1.25 with headers map[Authorization:[AAAAC3NzaC1lZDI1NTE5AAAAIAhd0mgYr2Vcgo+erwm80qQblGZaB2w4wE8iKVd5CEFx:dwydcT5JCB5WxCq5C7VD6Six77HhMVstFn6HMCBxU54YhtiEEYsOcHWMwC4EDADcM5NZD4oTgKf3/naoGa50CA==] User-Agent:[ollama/0.1.25 (amd64 windows) Go/go1.22.0]]" time=2024-02-21T10:37:46.246+01:00 level=DEBUG source=updater.go:96 msg="check update response 204 (current version is up to date)" time=2024-02-21T10:37:46.982+01:00 level=WARN source=server.go:109 msg="server crash 3 - exit code 1 - respawning" time=2024-02-21T10:37:47.493+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" time=2024-02-21T10:37:50.502+01:00 level=WARN source=server.go:109 msg="server crash 4 - exit code 1 - respawning" time=2024-02-21T10:37:51.016+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" time=2024-02-21T10:37:55.030+01:00 level=WARN source=server.go:109 msg="server crash 5 - exit code 1 - respawning" time=2024-02-21T10:37:55.543+01:00 level=ERROR source=server.go:112 msg="failed to restart server exec: already started" ``` **server.log** `Error: listen tcp 127.0.0.1:11434: bind: An attempt was made to access a socket in a way forbidden by its access permissions.` What I tried: [Ollama issue 2560](https://github.com/ollama/ollama/issues/2560#issuecomment-1950690705), but seems that there is no proxy at that port
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2627/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2627/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2624
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2624/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2624/comments
https://api.github.com/repos/ollama/ollama/issues/2624/events
https://github.com/ollama/ollama/issues/2624
2,145,890,647
I_kwDOJ0Z1Ps5_57FX
2,624
Support for Tinyllava
{ "login": "oliverbob", "id": 23272429, "node_id": "MDQ6VXNlcjIzMjcyNDI5", "avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4", "gravatar_id": "", "url": "https://api.github.com/users/oliverbob", "html_url": "https://github.com/oliverbob", "followers_url": "https://api.github.com/users/oliverbob/followers", "following_url": "https://api.github.com/users/oliverbob/following{/other_user}", "gists_url": "https://api.github.com/users/oliverbob/gists{/gist_id}", "starred_url": "https://api.github.com/users/oliverbob/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/oliverbob/subscriptions", "organizations_url": "https://api.github.com/users/oliverbob/orgs", "repos_url": "https://api.github.com/users/oliverbob/repos", "events_url": "https://api.github.com/users/oliverbob/events{/privacy}", "received_events_url": "https://api.github.com/users/oliverbob/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
3
2024-02-21T05:49:08
2024-09-06T14:39:07
2024-05-11T00:40:50
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
In addition to support for moondream #2259 which is lightning faster than llava, I believe there are some good guys there that can successfully make a GGUF version of Tinyllava which is faster than light. Tried it on safetensors to work really well. its been trained on the llava dataset, just like moondream. https://huggingface.co/bczhou/tiny-llava-v1-hf https://huggingface.co/YouLiXiya/tinyllava-v1.0-1.1b-hf
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2624/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 2 }
https://api.github.com/repos/ollama/ollama/issues/2624/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1867
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1867/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1867/comments
https://api.github.com/repos/ollama/ollama/issues/1867/events
https://github.com/ollama/ollama/issues/1867
2,072,415,115
I_kwDOJ0Z1Ps57houL
1,867
ollama barely uses any Ram
{ "login": "neuleo", "id": 99101285, "node_id": "U_kgDOBegqZQ", "avatar_url": "https://avatars.githubusercontent.com/u/99101285?v=4", "gravatar_id": "", "url": "https://api.github.com/users/neuleo", "html_url": "https://github.com/neuleo", "followers_url": "https://api.github.com/users/neuleo/followers", "following_url": "https://api.github.com/users/neuleo/following{/other_user}", "gists_url": "https://api.github.com/users/neuleo/gists{/gist_id}", "starred_url": "https://api.github.com/users/neuleo/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/neuleo/subscriptions", "organizations_url": "https://api.github.com/users/neuleo/orgs", "repos_url": "https://api.github.com/users/neuleo/repos", "events_url": "https://api.github.com/users/neuleo/events{/privacy}", "received_events_url": "https://api.github.com/users/neuleo/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2024-01-09T13:54:13
2024-01-09T18:43:22
2024-01-09T18:43:22
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hey Guys, I run ollama on docker and use mostly 7b models. But my Ram usage stays under 4 GB. Sometimes even below 3 GB. But the recommendations are 8 GB of Ram. It has 4 Core CPU, and it generates very slow even though I got 24 GB of Ram. I don't have a Video Card, though. I'm new to this, so can anyone tell me what I might need to do differently?
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1867/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1867/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2516
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2516/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2516/comments
https://api.github.com/repos/ollama/ollama/issues/2516/events
https://github.com/ollama/ollama/pull/2516
2,136,985,321
PR_kwDOJ0Z1Ps5nAGCZ
2,516
Fix a couple duplicate instance bugs
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-02-15T16:41:42
2024-02-15T23:52:46
2024-02-15T23:52:43
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2516", "html_url": "https://github.com/ollama/ollama/pull/2516", "diff_url": "https://github.com/ollama/ollama/pull/2516.diff", "patch_url": "https://github.com/ollama/ollama/pull/2516.patch", "merged_at": "2024-02-15T23:52:43" }
- Prevent the installer running multiple times concurrently - Detect multiple apps running and exit ~~with a message~~
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2516/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2516/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8255
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8255/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8255/comments
https://api.github.com/repos/ollama/ollama/issues/8255/events
https://github.com/ollama/ollama/issues/8255
2,760,642,962
I_kwDOJ0Z1Ps6kjBGS
8,255
question: the Windows version is very slow when accessing the API
{ "login": "liu9187", "id": 38241603, "node_id": "MDQ6VXNlcjM4MjQxNjAz", "avatar_url": "https://avatars.githubusercontent.com/u/38241603?v=4", "gravatar_id": "", "url": "https://api.github.com/users/liu9187", "html_url": "https://github.com/liu9187", "followers_url": "https://api.github.com/users/liu9187/followers", "following_url": "https://api.github.com/users/liu9187/following{/other_user}", "gists_url": "https://api.github.com/users/liu9187/gists{/gist_id}", "starred_url": "https://api.github.com/users/liu9187/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/liu9187/subscriptions", "organizations_url": "https://api.github.com/users/liu9187/orgs", "repos_url": "https://api.github.com/users/liu9187/repos", "events_url": "https://api.github.com/users/liu9187/events{/privacy}", "received_events_url": "https://api.github.com/users/liu9187/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5808482718, "node_id": "LA_kwDOJ0Z1Ps8AAAABWjZpng", "url": "https://api.github.com/repos/ollama/ollama/labels/performance", "name": "performance", "color": "A5B5C6", "default": false, "description": "" }, { "id": 6677367769, "node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q", "url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info", "name": "needs more info", "color": "BA8041", "default": false, "description": "More information is needed to assist" } ]
closed
false
null
[]
null
7
2024-12-27T09:34:12
2024-12-30T02:38:24
2024-12-30T02:38:24
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Excuse me, the Windows version is very slow when accessing the API. What is the reason? But using the command line is faster system : windows memory : 24G
{ "login": "liu9187", "id": 38241603, "node_id": "MDQ6VXNlcjM4MjQxNjAz", "avatar_url": "https://avatars.githubusercontent.com/u/38241603?v=4", "gravatar_id": "", "url": "https://api.github.com/users/liu9187", "html_url": "https://github.com/liu9187", "followers_url": "https://api.github.com/users/liu9187/followers", "following_url": "https://api.github.com/users/liu9187/following{/other_user}", "gists_url": "https://api.github.com/users/liu9187/gists{/gist_id}", "starred_url": "https://api.github.com/users/liu9187/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/liu9187/subscriptions", "organizations_url": "https://api.github.com/users/liu9187/orgs", "repos_url": "https://api.github.com/users/liu9187/repos", "events_url": "https://api.github.com/users/liu9187/events{/privacy}", "received_events_url": "https://api.github.com/users/liu9187/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8255/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8255/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/4462
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4462/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4462/comments
https://api.github.com/repos/ollama/ollama/issues/4462/events
https://github.com/ollama/ollama/pull/4462
2,298,992,426
PR_kwDOJ0Z1Ps5vl1FP
4,462
Port cuda/rocm skip build vars to linux
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-05-15T22:59:12
2024-05-15T23:27:50
2024-05-15T23:27:47
COLLABORATOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/4462", "html_url": "https://github.com/ollama/ollama/pull/4462", "diff_url": "https://github.com/ollama/ollama/pull/4462.diff", "patch_url": "https://github.com/ollama/ollama/pull/4462.patch", "merged_at": "2024-05-15T23:27:47" }
Windows already implements these, carry over to linux.
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4462/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4462/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/1141
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1141/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1141/comments
https://api.github.com/repos/ollama/ollama/issues/1141/events
https://github.com/ollama/ollama/issues/1141
1,995,322,582
I_kwDOJ0Z1Ps527jTW
1,141
Support forvietnamese-llama2-7b
{ "login": "khoint0210", "id": 12799726, "node_id": "MDQ6VXNlcjEyNzk5NzI2", "avatar_url": "https://avatars.githubusercontent.com/u/12799726?v=4", "gravatar_id": "", "url": "https://api.github.com/users/khoint0210", "html_url": "https://github.com/khoint0210", "followers_url": "https://api.github.com/users/khoint0210/followers", "following_url": "https://api.github.com/users/khoint0210/following{/other_user}", "gists_url": "https://api.github.com/users/khoint0210/gists{/gist_id}", "starred_url": "https://api.github.com/users/khoint0210/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/khoint0210/subscriptions", "organizations_url": "https://api.github.com/users/khoint0210/orgs", "repos_url": "https://api.github.com/users/khoint0210/repos", "events_url": "https://api.github.com/users/khoint0210/events{/privacy}", "received_events_url": "https://api.github.com/users/khoint0210/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2023-11-15T18:38:50
2024-12-23T01:10:18
2024-12-23T01:10:17
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi folk Really love this project I just wonder can you guy support this model https://huggingface.co/bkai-foundation-models/vietnamese-llama2-7b-40GB It would be fantastic to have this run inside Ollama
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1141/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1141/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2620
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2620/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2620/comments
https://api.github.com/repos/ollama/ollama/issues/2620/events
https://github.com/ollama/ollama/issues/2620
2,145,572,225
I_kwDOJ0Z1Ps5_4tWB
2,620
[Thankyou] Thanks for the ollama community
{ "login": "chuangtc", "id": 2288469, "node_id": "MDQ6VXNlcjIyODg0Njk=", "avatar_url": "https://avatars.githubusercontent.com/u/2288469?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chuangtc", "html_url": "https://github.com/chuangtc", "followers_url": "https://api.github.com/users/chuangtc/followers", "following_url": "https://api.github.com/users/chuangtc/following{/other_user}", "gists_url": "https://api.github.com/users/chuangtc/gists{/gist_id}", "starred_url": "https://api.github.com/users/chuangtc/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/chuangtc/subscriptions", "organizations_url": "https://api.github.com/users/chuangtc/orgs", "repos_url": "https://api.github.com/users/chuangtc/repos", "events_url": "https://api.github.com/users/chuangtc/events{/privacy}", "received_events_url": "https://api.github.com/users/chuangtc/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-02-21T01:06:32
2024-03-11T21:09:58
2024-03-11T21:09:58
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
In Taiwanese community, we have done a survey to see which are the top 5 local LLMs used by AI enthusiasts. ollama is list as the top 5. Here are the candidate list. https://matilabs.ai/2024/02/07/run-llms-locally/ Thanks for the ollama community's hard work. We really love this project. I feel it's easy and straight forward. Developer friendly. <img width="412" alt="Local-LLMs-2024-02-21 085848" src="https://github.com/ollama/ollama/assets/2288469/f8534a53-f4b3-439b-a2ba-2fa37e783bee">
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2620/reactions", "total_count": 5, "+1": 5, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2620/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3510
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3510/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3510/comments
https://api.github.com/repos/ollama/ollama/issues/3510/events
https://github.com/ollama/ollama/issues/3510
2,229,108,379
I_kwDOJ0Z1Ps6E3X6b
3,510
Databases
{ "login": "trymeouteh", "id": 31172274, "node_id": "MDQ6VXNlcjMxMTcyMjc0", "avatar_url": "https://avatars.githubusercontent.com/u/31172274?v=4", "gravatar_id": "", "url": "https://api.github.com/users/trymeouteh", "html_url": "https://github.com/trymeouteh", "followers_url": "https://api.github.com/users/trymeouteh/followers", "following_url": "https://api.github.com/users/trymeouteh/following{/other_user}", "gists_url": "https://api.github.com/users/trymeouteh/gists{/gist_id}", "starred_url": "https://api.github.com/users/trymeouteh/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/trymeouteh/subscriptions", "organizations_url": "https://api.github.com/users/trymeouteh/orgs", "repos_url": "https://api.github.com/users/trymeouteh/repos", "events_url": "https://api.github.com/users/trymeouteh/events{/privacy}", "received_events_url": "https://api.github.com/users/trymeouteh/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
1
2024-04-06T06:42:46
2024-05-09T02:26:26
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What are you trying to do? A way to manage, and create and train databases that can run on-top of models ### How should we solve this? Docker has images, containers and volumes. Ollama currently only has models. To add a new category for databases. Databases can be managed by... - Import from file - Export to file - Download from registry - Uploaded to registry - Create new empty database - Rename database - Delete database - Duplicate database (Copy/paste database) Databases can be edited by training the databases on files (documents, images, videos, audios, web pages, etc). A Ollama GUI would be best for training a database. Now you have models and databases, when you run a model, you can add one or more databases with it which it will use as additional trained data. ### What is the impact of not solving this? Ollama will be limited to LLMs and unable to add additional training data on-top of LLMs. ### Anything else? Adding an additional feature when running models to choose a database to send any trained data from prompt. If a response is good, user can send a like (thumbs up) to the database to train the model that the response was good. If a response is bad, user can send a dislike (thumbs down) to the database to train the model that the response was bad. Adding a tagging feature to databases would be useful https://github.com/ollama/ollama/issues/2847
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3510/reactions", "total_count": 5, "+1": 2, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 1, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3510/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/8046
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8046/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8046/comments
https://api.github.com/repos/ollama/ollama/issues/8046/events
https://github.com/ollama/ollama/issues/8046
2,733,068,523
I_kwDOJ0Z1Ps6i51Dr
8,046
Toggle theme
{ "login": "Abubakkar13", "id": 45032674, "node_id": "MDQ6VXNlcjQ1MDMyNjc0", "avatar_url": "https://avatars.githubusercontent.com/u/45032674?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Abubakkar13", "html_url": "https://github.com/Abubakkar13", "followers_url": "https://api.github.com/users/Abubakkar13/followers", "following_url": "https://api.github.com/users/Abubakkar13/following{/other_user}", "gists_url": "https://api.github.com/users/Abubakkar13/gists{/gist_id}", "starred_url": "https://api.github.com/users/Abubakkar13/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Abubakkar13/subscriptions", "organizations_url": "https://api.github.com/users/Abubakkar13/orgs", "repos_url": "https://api.github.com/users/Abubakkar13/repos", "events_url": "https://api.github.com/users/Abubakkar13/events{/privacy}", "received_events_url": "https://api.github.com/users/Abubakkar13/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" }, { "id": 6573197867, "node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw", "url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com", "name": "ollama.com", "color": "ffffff", "default": false, "description": "" } ]
closed
false
{ "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "hoyyeva", "id": 63033505, "node_id": "MDQ6VXNlcjYzMDMzNTA1", "avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hoyyeva", "html_url": "https://github.com/hoyyeva", "followers_url": "https://api.github.com/users/hoyyeva/followers", "following_url": "https://api.github.com/users/hoyyeva/following{/other_user}", "gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}", "starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions", "organizations_url": "https://api.github.com/users/hoyyeva/orgs", "repos_url": "https://api.github.com/users/hoyyeva/repos", "events_url": "https://api.github.com/users/hoyyeva/events{/privacy}", "received_events_url": "https://api.github.com/users/hoyyeva/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2024-12-11T14:03:39
2024-12-25T21:51:26
2024-12-25T21:51:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hey, Could you please add a feature for toggling between dark and light modes to ollama site? It would greatly enhance user experience by allowing customization based on preference and environment. A simple toggle button would be ideal!
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8046/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8046/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8656
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8656/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8656/comments
https://api.github.com/repos/ollama/ollama/issues/8656/events
https://github.com/ollama/ollama/pull/8656
2,818,099,446
PR_kwDOJ0Z1Ps6JWxCt
8,656
Add DeepSeek R1 in README
{ "login": "zakk616", "id": 26119949, "node_id": "MDQ6VXNlcjI2MTE5OTQ5", "avatar_url": "https://avatars.githubusercontent.com/u/26119949?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zakk616", "html_url": "https://github.com/zakk616", "followers_url": "https://api.github.com/users/zakk616/followers", "following_url": "https://api.github.com/users/zakk616/following{/other_user}", "gists_url": "https://api.github.com/users/zakk616/gists{/gist_id}", "starred_url": "https://api.github.com/users/zakk616/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/zakk616/subscriptions", "organizations_url": "https://api.github.com/users/zakk616/orgs", "repos_url": "https://api.github.com/users/zakk616/repos", "events_url": "https://api.github.com/users/zakk616/events{/privacy}", "received_events_url": "https://api.github.com/users/zakk616/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
open
false
null
[]
null
2
2025-01-29T12:38:21
2025-01-30T05:37:47
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/8656", "html_url": "https://github.com/ollama/ollama/pull/8656", "diff_url": "https://github.com/ollama/ollama/pull/8656.diff", "patch_url": "https://github.com/ollama/ollama/pull/8656.patch", "merged_at": null }
null
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8656/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8656/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/3136
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3136/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3136/comments
https://api.github.com/repos/ollama/ollama/issues/3136/events
https://github.com/ollama/ollama/issues/3136
2,185,694,105
I_kwDOJ0Z1Ps6CRwuZ
3,136
How to install offline? Is there an installation package available for download
{ "login": "yuanjie-ai", "id": 20265321, "node_id": "MDQ6VXNlcjIwMjY1MzIx", "avatar_url": "https://avatars.githubusercontent.com/u/20265321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yuanjie-ai", "html_url": "https://github.com/yuanjie-ai", "followers_url": "https://api.github.com/users/yuanjie-ai/followers", "following_url": "https://api.github.com/users/yuanjie-ai/following{/other_user}", "gists_url": "https://api.github.com/users/yuanjie-ai/gists{/gist_id}", "starred_url": "https://api.github.com/users/yuanjie-ai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/yuanjie-ai/subscriptions", "organizations_url": "https://api.github.com/users/yuanjie-ai/orgs", "repos_url": "https://api.github.com/users/yuanjie-ai/repos", "events_url": "https://api.github.com/users/yuanjie-ai/events{/privacy}", "received_events_url": "https://api.github.com/users/yuanjie-ai/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" } ]
closed
false
null
[]
null
2
2024-03-14T08:05:29
2024-03-14T20:26:34
2024-03-14T20:25:28
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
How to install offline? Is there an installation package available for download
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3136/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3136/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6068
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6068/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6068/comments
https://api.github.com/repos/ollama/ollama/issues/6068/events
https://github.com/ollama/ollama/issues/6068
2,436,948,189
I_kwDOJ0Z1Ps6RQODd
6,068
ollama serve --choice a model name
{ "login": "ruanjianlun", "id": 146827112, "node_id": "U_kgDOCMBnaA", "avatar_url": "https://avatars.githubusercontent.com/u/146827112?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ruanjianlun", "html_url": "https://github.com/ruanjianlun", "followers_url": "https://api.github.com/users/ruanjianlun/followers", "following_url": "https://api.github.com/users/ruanjianlun/following{/other_user}", "gists_url": "https://api.github.com/users/ruanjianlun/gists{/gist_id}", "starred_url": "https://api.github.com/users/ruanjianlun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ruanjianlun/subscriptions", "organizations_url": "https://api.github.com/users/ruanjianlun/orgs", "repos_url": "https://api.github.com/users/ruanjianlun/repos", "events_url": "https://api.github.com/users/ruanjianlun/events{/privacy}", "received_events_url": "https://api.github.com/users/ruanjianlun/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-07-30T05:44:11
2024-07-31T01:39:26
2024-07-31T01:39:26
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
hi guys I have a problem with aggregation, I am using it on windows, when I pass ollama serve, how do I select the aggregation model? For example, the following operation (base) PS C:\Users\Administrator> ollama serve --model codellama:7b Error: unknown flag: --model (base) PS C:\Users\Administrator>
{ "login": "ruanjianlun", "id": 146827112, "node_id": "U_kgDOCMBnaA", "avatar_url": "https://avatars.githubusercontent.com/u/146827112?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ruanjianlun", "html_url": "https://github.com/ruanjianlun", "followers_url": "https://api.github.com/users/ruanjianlun/followers", "following_url": "https://api.github.com/users/ruanjianlun/following{/other_user}", "gists_url": "https://api.github.com/users/ruanjianlun/gists{/gist_id}", "starred_url": "https://api.github.com/users/ruanjianlun/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ruanjianlun/subscriptions", "organizations_url": "https://api.github.com/users/ruanjianlun/orgs", "repos_url": "https://api.github.com/users/ruanjianlun/repos", "events_url": "https://api.github.com/users/ruanjianlun/events{/privacy}", "received_events_url": "https://api.github.com/users/ruanjianlun/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6068/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6068/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8033
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8033/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8033/comments
https://api.github.com/repos/ollama/ollama/issues/8033/events
https://github.com/ollama/ollama/issues/8033
2,731,004,328
I_kwDOJ0Z1Ps6ix9Go
8,033
nvcc compilation problem -- error: user-defined literal operator not found
{ "login": "envolution", "id": 12188773, "node_id": "MDQ6VXNlcjEyMTg4Nzcz", "avatar_url": "https://avatars.githubusercontent.com/u/12188773?v=4", "gravatar_id": "", "url": "https://api.github.com/users/envolution", "html_url": "https://github.com/envolution", "followers_url": "https://api.github.com/users/envolution/followers", "following_url": "https://api.github.com/users/envolution/following{/other_user}", "gists_url": "https://api.github.com/users/envolution/gists{/gist_id}", "starred_url": "https://api.github.com/users/envolution/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/envolution/subscriptions", "organizations_url": "https://api.github.com/users/envolution/orgs", "repos_url": "https://api.github.com/users/envolution/repos", "events_url": "https://api.github.com/users/envolution/events{/privacy}", "received_events_url": "https://api.github.com/users/envolution/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" }, { "id": 7700262114, "node_id": "LA_kwDOJ0Z1Ps8AAAAByvis4g", "url": "https://api.github.com/repos/ollama/ollama/labels/build", "name": "build", "color": "006b75", "default": false, "description": "Issues relating to building ollama from source" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
8
2024-12-10T19:27:04
2025-01-10T21:35:32
2024-12-16T06:05:30
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Compilation fails during 'make cuda_v12' ### Environment ``` $ /opt/cuda/bin/nvcc --version nvcc: NVIDIA (R) Cuda compiler driver Copyright (c) 2005-2024 NVIDIA Corporation Built on Tue_Oct_29_23:50:19_PDT_2024 Cuda compilation tools, release 12.6, V12.6.85 Build cuda_12.6.r12.6/compiler.35059454_0 $ gcc --version gcc (GCC) 14.2.1 20240910 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 565.77 Driver Version: 565.77 CUDA Version: 12.7 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce GTX 980 Off | 00000000:01:00.0 Off | N/A | | 29% 31C P8 13W / 180W | 2MiB / 4096MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+ ``` # Log $ make cuda_v12 CUDA_12_PATH=/opt/cuda /usr/bin/ccache /opt/cuda/bin/nvcc -c -Xcompiler -fPIC -D_GNU_SOURCE -fPIC -Wno-unused-function -std=c++11 -Xcompiler "-mavx" -t2 -DGGML_CUDA_DMMV_X=32 -DGGML_CUDA_MMV_Y=1 -DGGML_CUDA_PEER_MAX_BATCH_SIZE=128 -DGGML_USE_CUDA=1 -DGGML_SHARED=1 -DGGML_BUILD=1 -DGGML_USE_LLAMAFILE -DK_QUANTS_PER_ITERATION=2 -DNDEBUG -D_GNU_SOURCE -D_XOPEN_SOURCE=600 -Wno-deprecated-gpu-targets --forward-unknown-to-host-compiler -use_fast_math -I./llama/ -O3 --generate-code=arch=compute_60,code=[compute_60,sm_60] --generate-code=arch=compute_61,code=[compute_61,sm_61] --generate-code=arch=compute_62,code=[compute_62,sm_62] --generate-code=arch=compute_70,code=[compute_70,sm_70] --generate-code=arch=compute_72,code=[compute_72,sm_72] --generate-code=arch=compute_75,code=[compute_75,sm_75] --generate-code=arch=compute_80,code=[compute_80,sm_80] --generate-code=arch=compute_86,code=[compute_86,sm_86] --generate-code=arch=compute_87,code=[compute_87,sm_87] --generate-code=arch=compute_89,code=[compute_89,sm_89] --generate-code=arch=compute_90,code=[compute_90,sm_90] --generate-code=arch=compute_90a,code=[compute_90a,sm_90a] -DGGML_CUDA_USE_GRAPHS=1 -o llama/build/linux-amd64/llama/ggml-cuda.cuda_v12.o llama/ggml-cuda.cu /usr/include/c++/14.2.1/x86_64-pc-linux-gnu/bits/c++config.h(827): error: user-defined literal operator not found typedef __decltype(0.0bf16) __bfloat16_t; ^ /usr/include/c++/14.2.1/type_traits(529): error: type name is not allowed : public __bool_constant<__is_array(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(529): error: identifier "__is_array" is undefined : public __bool_constant<__is_array(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(581): error: type name is not allowed : public __bool_constant<__is_member_object_pointer(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(581): error: identifier "__is_member_object_pointer" is undefined : public __bool_constant<__is_member_object_pointer(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(603): error: type name is not allowed : public __bool_constant<__is_member_function_pointer(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(603): error: identifier "__is_member_function_pointer" is undefined : public __bool_constant<__is_member_function_pointer(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(695): error: type name is not allowed : public __bool_constant<__is_reference(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(695): error: identifier "__is_reference" is undefined : public __bool_constant<__is_reference(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(731): error: type name is not allowed : public __bool_constant<__is_object(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(731): error: identifier "__is_object" is undefined : public __bool_constant<__is_object(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(760): error: type name is not allowed : public __bool_constant<__is_member_pointer(_Tp)> ^ /usr/include/c++/14.2.1/type_traits(760): error: identifier "__is_member_pointer" is undefined : public __bool_constant<__is_member_pointer(_Tp)> ^ /usr/include/c++/14.2.1/bits/utility.h(237): error: __type_pack_element is not a template { using type = __type_pack_element<_Np, _Types...>; }; ^ /usr/include/c++/14.2.1/type_traits(138): error: class "std::enable_if<<error-constant>, void>" has no member "type" using __enable_if_t = typename enable_if<_Cond, _Tp>::type; ^ detected during: instantiation of type "std::__enable_if_t<<error-constant>, void>" at line 176 instantiation of "std::__detail::__or_fn" based on template arguments <std::is_reference<std::allocator<char>>, std::is_function<std::allocator<char>>, std::is_void<std::allocator<char>>, std::__is_array_unknown_bounds<std::allocator<char>>> at line 194 instantiation of class "std::__or_<_Bn...> [with _Bn=<std::is_reference<std::allocator<char>>, std::is_function<std::allocator<char>>, std::is_void<std::allocator<char>>, std::__is_array_unknown_bounds<std::allocator<char>>>]" at line 1195 instantiation of class "std::is_nothrow_default_constructible<_Tp> [with _Tp=std::allocator<char>]" at line 528 of /usr/include/c++/14.2.1/bits/basic_string.h instantiation of "std::__cxx11::basic_string<_CharT, _Traits, _Alloc>::basic_string() [with _CharT=char, _Traits=std::char_traits<char>, _Alloc=std::allocator<char>]" at line 4248 of /usr/include/c++/14.2.1/bits/basic_string.h 15 errors detected in the compilation of "llama/ggml-cuda.cu". make[1]: *** [make/gpu.make:54: llama/build/linux-amd64/llama/ggml-cuda.cuda_v12.o] Error 255 make: *** [Makefile:50: cuda_v12] Error 2 ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version git clone
{ "login": "envolution", "id": 12188773, "node_id": "MDQ6VXNlcjEyMTg4Nzcz", "avatar_url": "https://avatars.githubusercontent.com/u/12188773?v=4", "gravatar_id": "", "url": "https://api.github.com/users/envolution", "html_url": "https://github.com/envolution", "followers_url": "https://api.github.com/users/envolution/followers", "following_url": "https://api.github.com/users/envolution/following{/other_user}", "gists_url": "https://api.github.com/users/envolution/gists{/gist_id}", "starred_url": "https://api.github.com/users/envolution/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/envolution/subscriptions", "organizations_url": "https://api.github.com/users/envolution/orgs", "repos_url": "https://api.github.com/users/envolution/repos", "events_url": "https://api.github.com/users/envolution/events{/privacy}", "received_events_url": "https://api.github.com/users/envolution/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8033/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8033/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/6771
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6771/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6771/comments
https://api.github.com/repos/ollama/ollama/issues/6771/events
https://github.com/ollama/ollama/issues/6771
2,521,544,057
I_kwDOJ0Z1Ps6WS7V5
6,771
Inconsistent Responses from Identical Models
{ "login": "wahidur028", "id": 127589724, "node_id": "U_kgDOB5rdXA", "avatar_url": "https://avatars.githubusercontent.com/u/127589724?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wahidur028", "html_url": "https://github.com/wahidur028", "followers_url": "https://api.github.com/users/wahidur028/followers", "following_url": "https://api.github.com/users/wahidur028/following{/other_user}", "gists_url": "https://api.github.com/users/wahidur028/gists{/gist_id}", "starred_url": "https://api.github.com/users/wahidur028/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/wahidur028/subscriptions", "organizations_url": "https://api.github.com/users/wahidur028/orgs", "repos_url": "https://api.github.com/users/wahidur028/repos", "events_url": "https://api.github.com/users/wahidur028/events{/privacy}", "received_events_url": "https://api.github.com/users/wahidur028/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396220, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA", "url": "https://api.github.com/repos/ollama/ollama/labels/question", "name": "question", "color": "d876e3", "default": true, "description": "General questions" }, { "id": 7706482389, "node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q", "url": "https://api.github.com/repos/ollama/ollama/labels/api", "name": "api", "color": "bfdadc", "default": false, "description": "" } ]
closed
false
null
[]
null
1
2024-09-12T07:11:03
2024-11-06T00:32:29
2024-11-06T00:32:23
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I am new to Ollama and have noticed that when I ask a query using Ollama, the model's responses are quite poor. However, if I ask the same query using https://www.llama2.ai/, I receive much better responses. Can anyone explain what might be causing this difference? What could I be doing wrong? ``` # ! pip install -q pyautogen # ! pip install -q 'litellm[proxy]' import requests import json # Define the payload payload = { "model": "llama3:8b", # Changed model to match the curl version "prompt": """How does a neural network work?""" , "format": "json", # To ensure response is in JSON format "stream": False # To disable streaming in the response } # Send the POST request to the API response = requests.post("http://127.0.0.1:11434/api/generate", json=payload) output = response.json() print(output) ``` `{'model': 'llama3:8b', 'created_at': '2024-09-12T06:26:59.001500912Z', 'response': '{ }\n\n ', 'done': True, 'done_reason': 'stop', 'context': [128006, 882, 128007, 271, 4438, 1587, 264, 30828, 4009, 990, 30, 128009, 128006, 78191, 128007, 271, 90, 557, 220], 'total_duration': 243709137, 'load_duration': 56054134, 'prompt_eval_count': 17, 'prompt_eval_duration': 39720000, 'eval_count': 4, 'eval_duration': 103353000}` **Ollama model response:** response': '{ }\n\n **URL: https://www.llama2.ai/** **Model: Meta LLAMA3-8B** **Response:** I'd be happy to explain how a neural network works! A neural network is a type of machine learning model inspired by the structure and function of the human brain. It's a complex system made up of many interconnected nodes or "neurons," which process and transmit information. Here's a simplified overview of how a neural network works: 1. **Data Input**: The neural network receives input data, which can be images, sound waves, text, or any other type of data that can be digitized. 2. **Neuron Layers**: The input data is fed into the first layer of neurons, called the input layer. Each neuron in this layer receives a set of input values and performs a calculation on them. 3. **Activation Functions**: The output of each neuron is then passed through an activation function, which determines whether the neuron should "fire" or not. Common activation functions include sigmoid, tanh, and ReLU (Rectified Linear Unit). 4. **Hidden Layers**: The output from the input layer is then passed to one or more hidden layers, which are made up of neurons that perform complex calculations on the input data. Each hidden layer can have multiple neurons, and each neuron in the layer receives input from the previous layer. 5. **Output Layer**: The output from the hidden layers is then passed to the output layer, which produces the final output of the neural network. 6. **Training**: During training, the neural network is presented with a large dataset of input-output pairs. The network adjusts the weights and biases of the connections between neurons to minimize the error between its predictions and the actual output. 7. **Forward Propagation**: During forward propagation, the input data flows through the network, and the output is calculated at each layer. 8. **Backpropagation**: During backpropagation, the error between the predicted output and the actual output is calculated, and the error is propagated backwards through the network. This process helps the network adjust the weights and biases to improve its performance. 9. **Optimization**: The network uses an optimization algorithm, such as stochastic gradient descent (SGD), to adjust the weights and biases based on the error calculated during backpropagation. 10. **Repeat**: The process of forward propagation, backpropagation, and optimization is repeated multiple times until the network converges or reaches a desired level of accuracy. Neural networks can be used for a wide range of applications, including image and speech recognition, natural language processing, and predictive modeling. They're particularly useful when dealing with complex, non-linear relationships between inputs and outputs. I hope this helps! Do you have any specific questions about neural networks or would you like me to elaborate on any of these points? ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version _No response_
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6771/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6771/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/570
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/570/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/570/comments
https://api.github.com/repos/ollama/ollama/issues/570/events
https://github.com/ollama/ollama/pull/570
1,907,923,348
PR_kwDOJ0Z1Ps5a8JIo
570
fix HEAD request
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-09-21T23:41:17
2023-09-21T23:56:18
2023-09-21T23:56:17
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/570", "html_url": "https://github.com/ollama/ollama/pull/570", "diff_url": "https://github.com/ollama/ollama/pull/570.diff", "patch_url": "https://github.com/ollama/ollama/pull/570.patch", "merged_at": "2023-09-21T23:56:17" }
HEAD request should respond like their GET counterparts except without a response body. The previous implementation didn't quite satisfy this since it doesn't attach a response body so the content length is zero while the GET request responded with `Ollama is running`, content length 17. Despite having a response body, gin will omit it in the actual response: ``` $ http HEAD :11434/ HTTP/1.1 200 OK Content-Length: 17 Content-Type: text/plain; charset=utf-8 Date: Thu, 21 Sep 2023 23:41:08 GMT $ http GET :11434/ HTTP/1.1 200 OK Content-Length: 17 Content-Type: text/plain; charset=utf-8 Date: Thu, 21 Sep 2023 23:41:11 GMT Ollama is running ```
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/570/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/570/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/4511
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/4511/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/4511/comments
https://api.github.com/repos/ollama/ollama/issues/4511/events
https://github.com/ollama/ollama/issues/4511
2,303,962,194
I_kwDOJ0Z1Ps6JU6xS
4,511
Feature Request: Force-Off ROCm and CUDA builds in `gen_linux.sh` even if they are present.
{ "login": "dreirund", "id": 1590519, "node_id": "MDQ6VXNlcjE1OTA1MTk=", "avatar_url": "https://avatars.githubusercontent.com/u/1590519?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dreirund", "html_url": "https://github.com/dreirund", "followers_url": "https://api.github.com/users/dreirund/followers", "following_url": "https://api.github.com/users/dreirund/following{/other_user}", "gists_url": "https://api.github.com/users/dreirund/gists{/gist_id}", "starred_url": "https://api.github.com/users/dreirund/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dreirund/subscriptions", "organizations_url": "https://api.github.com/users/dreirund/orgs", "repos_url": "https://api.github.com/users/dreirund/repos", "events_url": "https://api.github.com/users/dreirund/events{/privacy}", "received_events_url": "https://api.github.com/users/dreirund/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
3
2024-05-18T09:38:09
2024-06-07T15:30:10
2024-05-21T20:24:55
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Ahoj, there is a need to force a build without CUDA or ROCm, even if some of their libraries are present on the system. But current `gen_linux.sh` forces ROCm build if some libraries are found, even if ROCm is not desired. [Arch Linux AUR package `ollama-nogpu-git`](https://aur.archlinux.org/pkgbase/ollama-nogpu-git) has a [patch to patch out all CUDA and ROCm stuff in `gen_linux.sh`](https://aur.archlinux.org/cgit/aur.git/tree/disable-rocm-cuda.gen_linux.sh.patch?h=ollama-nogpu-git), but [this upstream here seems to change things too fast that it is not feasible to stay up to date with that patch](https://aur.archlinux.org/pkgbase/ollama-nogpu-git#comment-973477). So I here wish for a feature to make it built-in that ROCm (and CUDA) build can be forced-off (e.g. by an environment variable or command line switch) _even if their libraries or just some of them_ are present. (On my system, I have only some stuff in `/opt/rocm/`, which is needed by some package, but is not a full and not a functional ROCm installation. This currently breaks building `ollama`.) Regards!
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/4511/reactions", "total_count": 3, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/4511/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/296
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/296/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/296/comments
https://api.github.com/repos/ollama/ollama/issues/296/events
https://github.com/ollama/ollama/issues/296
1,838,055,021
I_kwDOJ0Z1Ps5tjn5t
296
Provide a way to override system prompt at runtime
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
0
2023-08-06T05:16:49
2023-08-08T04:56:23
2023-08-08T04:56:23
MEMBER
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
```bash curl -X POST http://localhost:11434/api/generate -d '{ "model": "llama2", "system": "You are a helpful assistant.", "prompt": "hello" }' ```
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/296/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/296/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/3100
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/3100/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/3100/comments
https://api.github.com/repos/ollama/ollama/issues/3100/events
https://github.com/ollama/ollama/issues/3100
2,183,638,447
I_kwDOJ0Z1Ps6CJ62v
3,100
C4AI Command
{ "login": "AdaptiveStep", "id": 39104384, "node_id": "MDQ6VXNlcjM5MTA0Mzg0", "avatar_url": "https://avatars.githubusercontent.com/u/39104384?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AdaptiveStep", "html_url": "https://github.com/AdaptiveStep", "followers_url": "https://api.github.com/users/AdaptiveStep/followers", "following_url": "https://api.github.com/users/AdaptiveStep/following{/other_user}", "gists_url": "https://api.github.com/users/AdaptiveStep/gists{/gist_id}", "starred_url": "https://api.github.com/users/AdaptiveStep/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AdaptiveStep/subscriptions", "organizations_url": "https://api.github.com/users/AdaptiveStep/orgs", "repos_url": "https://api.github.com/users/AdaptiveStep/repos", "events_url": "https://api.github.com/users/AdaptiveStep/events{/privacy}", "received_events_url": "https://api.github.com/users/AdaptiveStep/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
4
2024-03-13T10:27:24
2024-04-15T15:36:36
2024-04-15T15:36:36
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Please add the c4ai-command model Its really good at translation and can handle 100 languages. https://huggingface.co/CohereForAI/c4ai-command-r-v01
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/3100/reactions", "total_count": 8, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/3100/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5203
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5203/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5203/comments
https://api.github.com/repos/ollama/ollama/issues/5203/events
https://github.com/ollama/ollama/issues/5203
2,367,053,249
I_kwDOJ0Z1Ps6NFl3B
5,203
OLLAMA_MODELS is not honored after being changed (beyond the first time it is set)
{ "login": "Nantris", "id": 6835891, "node_id": "MDQ6VXNlcjY4MzU4OTE=", "avatar_url": "https://avatars.githubusercontent.com/u/6835891?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nantris", "html_url": "https://github.com/Nantris", "followers_url": "https://api.github.com/users/Nantris/followers", "following_url": "https://api.github.com/users/Nantris/following{/other_user}", "gists_url": "https://api.github.com/users/Nantris/gists{/gist_id}", "starred_url": "https://api.github.com/users/Nantris/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nantris/subscriptions", "organizations_url": "https://api.github.com/users/Nantris/orgs", "repos_url": "https://api.github.com/users/Nantris/repos", "events_url": "https://api.github.com/users/Nantris/events{/privacy}", "received_events_url": "https://api.github.com/users/Nantris/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
open
false
null
[]
null
3
2024-06-21T18:28:08
2024-06-21T22:42:32
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? I set `OLLAMA_MODELS` to a new directory to ensure a large download would have space, and then tried to move `OLLAMA_MODELS` back to the default directory on Windows - and although the environmental variable is set, ollama continues to search in the previous `OLLAMA_MODELS` path, even after restarting all terminals or rebooting. ### OS Windows (not via WSL) ### GPU Nvidia ### CPU AMD ### Ollama version 0.1.44
{ "login": "Nantris", "id": 6835891, "node_id": "MDQ6VXNlcjY4MzU4OTE=", "avatar_url": "https://avatars.githubusercontent.com/u/6835891?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Nantris", "html_url": "https://github.com/Nantris", "followers_url": "https://api.github.com/users/Nantris/followers", "following_url": "https://api.github.com/users/Nantris/following{/other_user}", "gists_url": "https://api.github.com/users/Nantris/gists{/gist_id}", "starred_url": "https://api.github.com/users/Nantris/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Nantris/subscriptions", "organizations_url": "https://api.github.com/users/Nantris/orgs", "repos_url": "https://api.github.com/users/Nantris/repos", "events_url": "https://api.github.com/users/Nantris/events{/privacy}", "received_events_url": "https://api.github.com/users/Nantris/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5203/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5203/timeline
null
reopened
false
https://api.github.com/repos/ollama/ollama/issues/603
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/603/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/603/comments
https://api.github.com/repos/ollama/ollama/issues/603/events
https://github.com/ollama/ollama/issues/603
1,913,187,362
I_kwDOJ0Z1Ps5yCOwi
603
Considering graphql instead of classic http
{ "login": "FairyTail2000", "id": 22645621, "node_id": "MDQ6VXNlcjIyNjQ1NjIx", "avatar_url": "https://avatars.githubusercontent.com/u/22645621?v=4", "gravatar_id": "", "url": "https://api.github.com/users/FairyTail2000", "html_url": "https://github.com/FairyTail2000", "followers_url": "https://api.github.com/users/FairyTail2000/followers", "following_url": "https://api.github.com/users/FairyTail2000/following{/other_user}", "gists_url": "https://api.github.com/users/FairyTail2000/gists{/gist_id}", "starred_url": "https://api.github.com/users/FairyTail2000/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/FairyTail2000/subscriptions", "organizations_url": "https://api.github.com/users/FairyTail2000/orgs", "repos_url": "https://api.github.com/users/FairyTail2000/repos", "events_url": "https://api.github.com/users/FairyTail2000/events{/privacy}", "received_events_url": "https://api.github.com/users/FairyTail2000/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
2
2023-09-26T10:19:39
2023-09-30T05:09:45
2023-09-30T05:09:45
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I won't explain here what [graphql](https://graphql.org) is. ## How does this project benefit from graphql vs classical http? With graphql you can get more info / less info per http call. Why is this relevant? Third party integration. Example with my own frontend: - I want to get all models (1 call) - I want to inspect all models (x calls) I have right now 20 - 30 models and those I would make models + 1 call to the backend. Depending on the network conditions this would significantly impact loading time of the third party integration. Also I would always fetch all info, which isn't always wanted. Also I would always fetch "modified_at" for the all models, and for the model inspection I would always fetch "modelfile", "parameters" and "template". This increases the network payload size significantly How does graphql help with this? With graphql I can request a list of all models and related fields in a single field. Example: ```graphql query getAllModels { models { name size details { license } } } ``` With this I can make a single http request and can get all the info I need and want. Another plus is that there is a code generator available for many languages and framework. For the ollama cli client this api doesn't make much sense. This is a fundamental decision for the project and thus requires much thought before making a decision. I'm looking forward to feedback on this
{ "login": "mchiang0610", "id": 3325447, "node_id": "MDQ6VXNlcjMzMjU0NDc=", "avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mchiang0610", "html_url": "https://github.com/mchiang0610", "followers_url": "https://api.github.com/users/mchiang0610/followers", "following_url": "https://api.github.com/users/mchiang0610/following{/other_user}", "gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}", "starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions", "organizations_url": "https://api.github.com/users/mchiang0610/orgs", "repos_url": "https://api.github.com/users/mchiang0610/repos", "events_url": "https://api.github.com/users/mchiang0610/events{/privacy}", "received_events_url": "https://api.github.com/users/mchiang0610/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/603/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/603/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8450
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8450/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8450/comments
https://api.github.com/repos/ollama/ollama/issues/8450/events
https://github.com/ollama/ollama/issues/8450
2,792,010,100
I_kwDOJ0Z1Ps6marF0
8,450
ollama v0.5.6 /save bug?
{ "login": "Feng-Yong-Qi", "id": 130546218, "node_id": "U_kgDOB8f6Kg", "avatar_url": "https://avatars.githubusercontent.com/u/130546218?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Feng-Yong-Qi", "html_url": "https://github.com/Feng-Yong-Qi", "followers_url": "https://api.github.com/users/Feng-Yong-Qi/followers", "following_url": "https://api.github.com/users/Feng-Yong-Qi/following{/other_user}", "gists_url": "https://api.github.com/users/Feng-Yong-Qi/gists{/gist_id}", "starred_url": "https://api.github.com/users/Feng-Yong-Qi/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Feng-Yong-Qi/subscriptions", "organizations_url": "https://api.github.com/users/Feng-Yong-Qi/orgs", "repos_url": "https://api.github.com/users/Feng-Yong-Qi/repos", "events_url": "https://api.github.com/users/Feng-Yong-Qi/events{/privacy}", "received_events_url": "https://api.github.com/users/Feng-Yong-Qi/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
2
2025-01-16T08:23:31
2025-01-18T05:25:03
2025-01-18T05:25:03
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
### What is the issue? Excuse me, v0.5.6 version /save error. Operating System: Rocky Linux 9.4. The CPU architecture is x86. v0.5.5 VS v0.5.6 ![Image](https://github.com/user-attachments/assets/da79a935-d802-4714-9678-b37795cd2835) ![Image](https://github.com/user-attachments/assets/032cffca-a200-4e33-8d77-971d419cd3f6) ### OS Linux ### GPU Intel ### CPU Intel ### Ollama version 0.5.6
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8450/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8450/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8430
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8430/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8430/comments
https://api.github.com/repos/ollama/ollama/issues/8430/events
https://github.com/ollama/ollama/issues/8430
2,788,596,735
I_kwDOJ0Z1Ps6mNpv_
8,430
Multi GPU, default GPU setting, specific model pin to specific GPU
{ "login": "dariuszsekula", "id": 118318131, "node_id": "U_kgDOBw1kMw", "avatar_url": "https://avatars.githubusercontent.com/u/118318131?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dariuszsekula", "html_url": "https://github.com/dariuszsekula", "followers_url": "https://api.github.com/users/dariuszsekula/followers", "following_url": "https://api.github.com/users/dariuszsekula/following{/other_user}", "gists_url": "https://api.github.com/users/dariuszsekula/gists{/gist_id}", "starred_url": "https://api.github.com/users/dariuszsekula/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dariuszsekula/subscriptions", "organizations_url": "https://api.github.com/users/dariuszsekula/orgs", "repos_url": "https://api.github.com/users/dariuszsekula/repos", "events_url": "https://api.github.com/users/dariuszsekula/events{/privacy}", "received_events_url": "https://api.github.com/users/dariuszsekula/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
null
[]
null
4
2025-01-15T01:08:46
2025-01-16T18:05:05
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have a multi GPU configuration and they are different GPU models with different memory size. I wish: 1. we could select default GPU for all models (potentially the fastest one with higher memory) 2. we could select specific model per specific GPU to use small models into small VRAM GPU and use fastest/highest_VRAM card for other large models.
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8430/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8430/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7701
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7701/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7701/comments
https://api.github.com/repos/ollama/ollama/issues/7701/events
https://github.com/ollama/ollama/issues/7701
2,664,440,219
I_kwDOJ0Z1Ps6e0CGb
7,701
[question] Why are there such large differences in installation package sizes for different CPU architectures and systems?
{ "login": "netjune", "id": 17109782, "node_id": "MDQ6VXNlcjE3MTA5Nzgy", "avatar_url": "https://avatars.githubusercontent.com/u/17109782?v=4", "gravatar_id": "", "url": "https://api.github.com/users/netjune", "html_url": "https://github.com/netjune", "followers_url": "https://api.github.com/users/netjune/followers", "following_url": "https://api.github.com/users/netjune/following{/other_user}", "gists_url": "https://api.github.com/users/netjune/gists{/gist_id}", "starred_url": "https://api.github.com/users/netjune/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/netjune/subscriptions", "organizations_url": "https://api.github.com/users/netjune/orgs", "repos_url": "https://api.github.com/users/netjune/repos", "events_url": "https://api.github.com/users/netjune/events{/privacy}", "received_events_url": "https://api.github.com/users/netjune/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
0
2024-11-16T14:31:49
2024-11-17T09:41:56
2024-11-17T09:41:56
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
![screenshot-20241116-222333](https://github.com/user-attachments/assets/a649b559-5b4b-4b06-99e9-67ff7929606b) More than 10x
{ "login": "netjune", "id": 17109782, "node_id": "MDQ6VXNlcjE3MTA5Nzgy", "avatar_url": "https://avatars.githubusercontent.com/u/17109782?v=4", "gravatar_id": "", "url": "https://api.github.com/users/netjune", "html_url": "https://github.com/netjune", "followers_url": "https://api.github.com/users/netjune/followers", "following_url": "https://api.github.com/users/netjune/following{/other_user}", "gists_url": "https://api.github.com/users/netjune/gists{/gist_id}", "starred_url": "https://api.github.com/users/netjune/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/netjune/subscriptions", "organizations_url": "https://api.github.com/users/netjune/orgs", "repos_url": "https://api.github.com/users/netjune/repos", "events_url": "https://api.github.com/users/netjune/events{/privacy}", "received_events_url": "https://api.github.com/users/netjune/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7701/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7701/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/1206
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/1206/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/1206/comments
https://api.github.com/repos/ollama/ollama/issues/1206/events
https://github.com/ollama/ollama/pull/1206
2,002,338,841
PR_kwDOJ0Z1Ps5f6uPA
1,206
README: link to LangChainGo for talking to ollama, with an example
{ "login": "eliben", "id": 1130906, "node_id": "MDQ6VXNlcjExMzA5MDY=", "avatar_url": "https://avatars.githubusercontent.com/u/1130906?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eliben", "html_url": "https://github.com/eliben", "followers_url": "https://api.github.com/users/eliben/followers", "following_url": "https://api.github.com/users/eliben/following{/other_user}", "gists_url": "https://api.github.com/users/eliben/gists{/gist_id}", "starred_url": "https://api.github.com/users/eliben/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eliben/subscriptions", "organizations_url": "https://api.github.com/users/eliben/orgs", "repos_url": "https://api.github.com/users/eliben/repos", "events_url": "https://api.github.com/users/eliben/events{/privacy}", "received_events_url": "https://api.github.com/users/eliben/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2023-11-20T14:33:01
2023-11-20T15:35:07
2023-11-20T15:35:07
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/1206", "html_url": "https://github.com/ollama/ollama/pull/1206", "diff_url": "https://github.com/ollama/ollama/pull/1206.diff", "patch_url": "https://github.com/ollama/ollama/pull/1206.patch", "merged_at": "2023-11-20T15:35:07" }
null
{ "login": "BruceMacD", "id": 5853428, "node_id": "MDQ6VXNlcjU4NTM0Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BruceMacD", "html_url": "https://github.com/BruceMacD", "followers_url": "https://api.github.com/users/BruceMacD/followers", "following_url": "https://api.github.com/users/BruceMacD/following{/other_user}", "gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}", "starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions", "organizations_url": "https://api.github.com/users/BruceMacD/orgs", "repos_url": "https://api.github.com/users/BruceMacD/repos", "events_url": "https://api.github.com/users/BruceMacD/events{/privacy}", "received_events_url": "https://api.github.com/users/BruceMacD/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/1206/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/1206/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/5616
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5616/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5616/comments
https://api.github.com/repos/ollama/ollama/issues/5616/events
https://github.com/ollama/ollama/pull/5616
2,401,847,462
PR_kwDOJ0Z1Ps51Bni3
5,616
fix: quant err message
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-10T22:30:02
2024-07-12T00:24:31
2024-07-12T00:24:29
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5616", "html_url": "https://github.com/ollama/ollama/pull/5616", "diff_url": "https://github.com/ollama/ollama/pull/5616.diff", "patch_url": "https://github.com/ollama/ollama/pull/5616.patch", "merged_at": "2024-07-12T00:24:29" }
`Error: quantization of this model is not supported by your version of Ollama. You may need to upgrade` Resolves: https://github.com/ollama/ollama/issues/5531
{ "login": "joshyan1", "id": 76125168, "node_id": "MDQ6VXNlcjc2MTI1MTY4", "avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joshyan1", "html_url": "https://github.com/joshyan1", "followers_url": "https://api.github.com/users/joshyan1/followers", "following_url": "https://api.github.com/users/joshyan1/following{/other_user}", "gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}", "starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions", "organizations_url": "https://api.github.com/users/joshyan1/orgs", "repos_url": "https://api.github.com/users/joshyan1/repos", "events_url": "https://api.github.com/users/joshyan1/events{/privacy}", "received_events_url": "https://api.github.com/users/joshyan1/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5616/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5616/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/6198
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/6198/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/6198/comments
https://api.github.com/repos/ollama/ollama/issues/6198/events
https://github.com/ollama/ollama/issues/6198
2,450,702,986
I_kwDOJ0Z1Ps6SEsKK
6,198
Request to Add JAIS 70B Model
{ "login": "umar052001", "id": 79453927, "node_id": "MDQ6VXNlcjc5NDUzOTI3", "avatar_url": "https://avatars.githubusercontent.com/u/79453927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/umar052001", "html_url": "https://github.com/umar052001", "followers_url": "https://api.github.com/users/umar052001/followers", "following_url": "https://api.github.com/users/umar052001/following{/other_user}", "gists_url": "https://api.github.com/users/umar052001/gists{/gist_id}", "starred_url": "https://api.github.com/users/umar052001/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/umar052001/subscriptions", "organizations_url": "https://api.github.com/users/umar052001/orgs", "repos_url": "https://api.github.com/users/umar052001/repos", "events_url": "https://api.github.com/users/umar052001/events{/privacy}", "received_events_url": "https://api.github.com/users/umar052001/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
closed
false
null
[]
null
5
2024-08-06T11:58:03
2024-10-22T13:32:44
2024-10-22T13:32:44
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
**Description:** I would like to suggest adding the JAIS 70B model to Ollama. The model is available on Hugging Face [here](https://huggingface.co/inceptionai/jais-adapted-70b) and offers significant advancements in Arabic-English natural language processing. It would be a valuable addition for users interested in Arabic-centric NLP applications. **Reason for Request:** The JAIS 70B model is specifically designed to handle complex Arabic-English tasks, and its availability on Ollama would allow for broader testing and application in various use cases. This addition would enhance the platform’s offerings for users working with Arabic and bilingual datasets. **Reference:** - [JAIS 70B on Hugging Face](https://huggingface.co/inceptionai/jais-adapted-70b) **Additional Information:** This model was recently launched by Inception under the G42 group and has been well-received for its large-scale capabilities and efficient training on a vast dataset, including 330 billion Arabic tokens. Thank you for considering this suggestion!
{ "login": "umar052001", "id": 79453927, "node_id": "MDQ6VXNlcjc5NDUzOTI3", "avatar_url": "https://avatars.githubusercontent.com/u/79453927?v=4", "gravatar_id": "", "url": "https://api.github.com/users/umar052001", "html_url": "https://github.com/umar052001", "followers_url": "https://api.github.com/users/umar052001/followers", "following_url": "https://api.github.com/users/umar052001/following{/other_user}", "gists_url": "https://api.github.com/users/umar052001/gists{/gist_id}", "starred_url": "https://api.github.com/users/umar052001/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/umar052001/subscriptions", "organizations_url": "https://api.github.com/users/umar052001/orgs", "repos_url": "https://api.github.com/users/umar052001/repos", "events_url": "https://api.github.com/users/umar052001/events{/privacy}", "received_events_url": "https://api.github.com/users/umar052001/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/6198/reactions", "total_count": 21, "+1": 6, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 8, "rocket": 7, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/6198/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2277
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2277/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2277/comments
https://api.github.com/repos/ollama/ollama/issues/2277/events
https://github.com/ollama/ollama/issues/2277
2,108,235,095
I_kwDOJ0Z1Ps59qR1X
2,277
How to set ROCR_VISIBLE_DEVICES to 0
{ "login": "meminens", "id": 42714627, "node_id": "MDQ6VXNlcjQyNzE0NjI3", "avatar_url": "https://avatars.githubusercontent.com/u/42714627?v=4", "gravatar_id": "", "url": "https://api.github.com/users/meminens", "html_url": "https://github.com/meminens", "followers_url": "https://api.github.com/users/meminens/followers", "following_url": "https://api.github.com/users/meminens/following{/other_user}", "gists_url": "https://api.github.com/users/meminens/gists{/gist_id}", "starred_url": "https://api.github.com/users/meminens/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/meminens/subscriptions", "organizations_url": "https://api.github.com/users/meminens/orgs", "repos_url": "https://api.github.com/users/meminens/repos", "events_url": "https://api.github.com/users/meminens/events{/privacy}", "received_events_url": "https://api.github.com/users/meminens/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
2
2024-01-30T16:10:19
2024-01-31T21:15:19
2024-01-31T21:15:19
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I have installed ollama (v0.1.22) and ROCm (v5.7.1) to Arch Linux via the following commands ``` pacman -S ollama rocm-hip-sdk rocm-opencl-sdk clblast systemctl daemon-reload systemctl enable ollama.service systemctl start ollama.service ``` and then run `ollama run mistral` Checking `htop` and `nvtop`, I see that only CPU is being used. Ollama log in `journalctl -u ollama` shows the following: ``` Searching for GPU management library libnvidia-ml.so Discovered GPU libraries: [] Searching for GPU management library librocm_smi64.so Discovered GPU libraries: [/opt/rocm/lib/librocm_smi64.so.5.0] Radeon GPU detected ROCm integrated GPU detected - ROCR_VISIBLE_DEVICES=1 ``` I believe the `ROCR_VISIBLE_DEVICES` parameter should be set to `0`. My system info is below: ``` -` misaligar@arch .o+` --------- `ooo/ OS: Arch Linux x86_64 `+oooo: Host: B650 AORUS ELITE AX `+oooooo: Kernel: 6.7.2-arch1-1 -+oooooo+: Uptime: 18 mins `/:-:++oooo+: Packages: 1065 (pacman), 7 (flatpak) `/++++/+++++++: Shell: bash 5.2.26 `/++++++++++++++: Resolution: 2560x1440 `/+++ooooooooooooo/` DE: Plasma 5.27.10 ./ooosssso++osssssso+` WM: kwin .oossssso-````/ossssss+` Theme: [Plasma], Breeze [GTK2/3] -osssssso. :ssssssso. Icons: kora [Plasma], kora [GTK2/3] :osssssss/ osssso+++. Terminal: konsole /ossssssss/ +ssssooo/- Terminal Font: Hack Nerd Font Mono 10 `/ossssso+/:- -:/+osssso+- CPU: AMD Ryzen 9 7900X (24) @ 5.733GHz `+sso+:-` `.-/+oso: GPU: AMD ATI 13:00.0 Raphael `++:. `-/+/ GPU: AMD ATI Radeon RX 7900 XT/7900 XTX .` `/ Memory: 4811MiB / 63438MiB ``` How can I tell ollama to use the external GPU?
{ "login": "dhiltgen", "id": 4033016, "node_id": "MDQ6VXNlcjQwMzMwMTY=", "avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4", "gravatar_id": "", "url": "https://api.github.com/users/dhiltgen", "html_url": "https://github.com/dhiltgen", "followers_url": "https://api.github.com/users/dhiltgen/followers", "following_url": "https://api.github.com/users/dhiltgen/following{/other_user}", "gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}", "starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions", "organizations_url": "https://api.github.com/users/dhiltgen/orgs", "repos_url": "https://api.github.com/users/dhiltgen/repos", "events_url": "https://api.github.com/users/dhiltgen/events{/privacy}", "received_events_url": "https://api.github.com/users/dhiltgen/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2277/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2277/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/7874
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7874/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7874/comments
https://api.github.com/repos/ollama/ollama/issues/7874/events
https://github.com/ollama/ollama/issues/7874
2,702,895,284
I_kwDOJ0Z1Ps6hGui0
7,874
Provide a slash command to clear screen.
{ "login": "231tr0n", "id": 56352048, "node_id": "MDQ6VXNlcjU2MzUyMDQ4", "avatar_url": "https://avatars.githubusercontent.com/u/56352048?v=4", "gravatar_id": "", "url": "https://api.github.com/users/231tr0n", "html_url": "https://github.com/231tr0n", "followers_url": "https://api.github.com/users/231tr0n/followers", "following_url": "https://api.github.com/users/231tr0n/following{/other_user}", "gists_url": "https://api.github.com/users/231tr0n/gists{/gist_id}", "starred_url": "https://api.github.com/users/231tr0n/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/231tr0n/subscriptions", "organizations_url": "https://api.github.com/users/231tr0n/orgs", "repos_url": "https://api.github.com/users/231tr0n/repos", "events_url": "https://api.github.com/users/231tr0n/events{/privacy}", "received_events_url": "https://api.github.com/users/231tr0n/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
2
2024-11-28T17:54:54
2024-11-28T17:58:59
2024-11-28T17:56:21
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Hi! Is it possible to provide a slash command to clear the screen of the terminal chat window which appears when running a model with ollama run?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7874/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7874/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/726
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/726/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/726/comments
https://api.github.com/repos/ollama/ollama/issues/726/events
https://github.com/ollama/ollama/issues/726
1,931,195,476
I_kwDOJ0Z1Ps5zG7RU
726
Currently create is recrating whole model , how to just update?
{ "login": "v3ss0n", "id": 419606, "node_id": "MDQ6VXNlcjQxOTYwNg==", "avatar_url": "https://avatars.githubusercontent.com/u/419606?v=4", "gravatar_id": "", "url": "https://api.github.com/users/v3ss0n", "html_url": "https://github.com/v3ss0n", "followers_url": "https://api.github.com/users/v3ss0n/followers", "following_url": "https://api.github.com/users/v3ss0n/following{/other_user}", "gists_url": "https://api.github.com/users/v3ss0n/gists{/gist_id}", "starred_url": "https://api.github.com/users/v3ss0n/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/v3ss0n/subscriptions", "organizations_url": "https://api.github.com/users/v3ss0n/orgs", "repos_url": "https://api.github.com/users/v3ss0n/repos", "events_url": "https://api.github.com/users/v3ss0n/events{/privacy}", "received_events_url": "https://api.github.com/users/v3ss0n/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396184, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA", "url": "https://api.github.com/repos/ollama/ollama/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
1
2023-10-07T05:18:26
2023-10-28T19:26:48
2023-10-28T19:26:48
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
I am tuning pramaeters nad i have to re run `ollama create name modelfile` . That remake the model . If there is no model file changes can i just update the parameters?
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/726/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/726/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/2926
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/2926/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/2926/comments
https://api.github.com/repos/ollama/ollama/issues/2926/events
https://github.com/ollama/ollama/pull/2926
2,167,878,027
PR_kwDOJ0Z1Ps5opVJh
2,926
refactor model parsing
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-03-04T22:12:49
2024-04-01T20:58:14
2024-04-01T20:58:13
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/2926", "html_url": "https://github.com/ollama/ollama/pull/2926", "diff_url": "https://github.com/ollama/ollama/pull/2926.diff", "patch_url": "https://github.com/ollama/ollama/pull/2926.patch", "merged_at": "2024-04-01T20:58:13" }
- refactor metadata
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/2926/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/2926/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/8496
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8496/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8496/comments
https://api.github.com/repos/ollama/ollama/issues/8496/events
https://github.com/ollama/ollama/issues/8496
2,798,197,635
I_kwDOJ0Z1Ps6myRuD
8,496
Requesting this new multimodal model.
{ "login": "Ramachandra-2k96", "id": 149596008, "node_id": "U_kgDOCOqnaA", "avatar_url": "https://avatars.githubusercontent.com/u/149596008?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Ramachandra-2k96", "html_url": "https://github.com/Ramachandra-2k96", "followers_url": "https://api.github.com/users/Ramachandra-2k96/followers", "following_url": "https://api.github.com/users/Ramachandra-2k96/following{/other_user}", "gists_url": "https://api.github.com/users/Ramachandra-2k96/gists{/gist_id}", "starred_url": "https://api.github.com/users/Ramachandra-2k96/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Ramachandra-2k96/subscriptions", "organizations_url": "https://api.github.com/users/Ramachandra-2k96/orgs", "repos_url": "https://api.github.com/users/Ramachandra-2k96/repos", "events_url": "https://api.github.com/users/Ramachandra-2k96/events{/privacy}", "received_events_url": "https://api.github.com/users/Ramachandra-2k96/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5789807732, "node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA", "url": "https://api.github.com/repos/ollama/ollama/labels/model%20request", "name": "model request", "color": "1E5DE6", "default": false, "description": "Model requests" } ]
open
false
null
[]
null
3
2025-01-20T05:06:01
2025-01-23T19:16:03
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Please add the openbmb/MiniCPM-o-2_6 Model. https://huggingface.co/openbmb/MiniCPM-o-2_6
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8496/reactions", "total_count": 9, "+1": 8, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/ollama/ollama/issues/8496/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/7832
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7832/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7832/comments
https://api.github.com/repos/ollama/ollama/issues/7832/events
https://github.com/ollama/ollama/pull/7832
2,692,423,898
PR_kwDOJ0Z1Ps6DGnAU
7,832
Update README.md
{ "login": "jake83741", "id": 125723241, "node_id": "U_kgDOB35iaQ", "avatar_url": "https://avatars.githubusercontent.com/u/125723241?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jake83741", "html_url": "https://github.com/jake83741", "followers_url": "https://api.github.com/users/jake83741/followers", "following_url": "https://api.github.com/users/jake83741/following{/other_user}", "gists_url": "https://api.github.com/users/jake83741/gists{/gist_id}", "starred_url": "https://api.github.com/users/jake83741/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jake83741/subscriptions", "organizations_url": "https://api.github.com/users/jake83741/orgs", "repos_url": "https://api.github.com/users/jake83741/repos", "events_url": "https://api.github.com/users/jake83741/events{/privacy}", "received_events_url": "https://api.github.com/users/jake83741/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
1
2024-11-25T22:20:16
2024-11-26T02:39:37
2024-11-26T01:56:30
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/7832", "html_url": "https://github.com/ollama/ollama/pull/7832", "diff_url": "https://github.com/ollama/ollama/pull/7832.diff", "patch_url": "https://github.com/ollama/ollama/pull/7832.patch", "merged_at": "2024-11-26T01:56:30" }
Hi, This is an update on the description of my Discord bot project - [vnc-lm](https://github.com/jake83741/vnc-lm) in the README. Thank you, Jake
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7832/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7832/timeline
null
null
true
https://api.github.com/repos/ollama/ollama/issues/7523
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/7523/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/7523/comments
https://api.github.com/repos/ollama/ollama/issues/7523/events
https://github.com/ollama/ollama/issues/7523
2,637,275,073
I_kwDOJ0Z1Ps6dMZ_B
7,523
[FEAT] Add ollama installation selection
{ "login": "rong-xiaoli", "id": 58361774, "node_id": "MDQ6VXNlcjU4MzYxNzc0", "avatar_url": "https://avatars.githubusercontent.com/u/58361774?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rong-xiaoli", "html_url": "https://github.com/rong-xiaoli", "followers_url": "https://api.github.com/users/rong-xiaoli/followers", "following_url": "https://api.github.com/users/rong-xiaoli/following{/other_user}", "gists_url": "https://api.github.com/users/rong-xiaoli/gists{/gist_id}", "starred_url": "https://api.github.com/users/rong-xiaoli/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rong-xiaoli/subscriptions", "organizations_url": "https://api.github.com/users/rong-xiaoli/orgs", "repos_url": "https://api.github.com/users/rong-xiaoli/repos", "events_url": "https://api.github.com/users/rong-xiaoli/events{/privacy}", "received_events_url": "https://api.github.com/users/rong-xiaoli/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
5
2024-11-06T07:33:58
2024-11-06T11:45:31
2024-11-06T11:45:31
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
My computer has only 46MB left on my `C:` and ollama installation program fails to install. I tried my best to get more space but I could not. It seems like the installation program does not support installation path selection, thus I suggest adding a path selection before installation.
{ "login": "rong-xiaoli", "id": 58361774, "node_id": "MDQ6VXNlcjU4MzYxNzc0", "avatar_url": "https://avatars.githubusercontent.com/u/58361774?v=4", "gravatar_id": "", "url": "https://api.github.com/users/rong-xiaoli", "html_url": "https://github.com/rong-xiaoli", "followers_url": "https://api.github.com/users/rong-xiaoli/followers", "following_url": "https://api.github.com/users/rong-xiaoli/following{/other_user}", "gists_url": "https://api.github.com/users/rong-xiaoli/gists{/gist_id}", "starred_url": "https://api.github.com/users/rong-xiaoli/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/rong-xiaoli/subscriptions", "organizations_url": "https://api.github.com/users/rong-xiaoli/orgs", "repos_url": "https://api.github.com/users/rong-xiaoli/repos", "events_url": "https://api.github.com/users/rong-xiaoli/events{/privacy}", "received_events_url": "https://api.github.com/users/rong-xiaoli/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/7523/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/7523/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/8288
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/8288/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/8288/comments
https://api.github.com/repos/ollama/ollama/issues/8288/events
https://github.com/ollama/ollama/issues/8288
2,765,953,153
I_kwDOJ0Z1Ps6k3RiB
8,288
Enable auto-save functionality via CLI flag
{ "login": "migueltorrescosta", "id": 6451658, "node_id": "MDQ6VXNlcjY0NTE2NTg=", "avatar_url": "https://avatars.githubusercontent.com/u/6451658?v=4", "gravatar_id": "", "url": "https://api.github.com/users/migueltorrescosta", "html_url": "https://github.com/migueltorrescosta", "followers_url": "https://api.github.com/users/migueltorrescosta/followers", "following_url": "https://api.github.com/users/migueltorrescosta/following{/other_user}", "gists_url": "https://api.github.com/users/migueltorrescosta/gists{/gist_id}", "starred_url": "https://api.github.com/users/migueltorrescosta/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/migueltorrescosta/subscriptions", "organizations_url": "https://api.github.com/users/migueltorrescosta/orgs", "repos_url": "https://api.github.com/users/migueltorrescosta/repos", "events_url": "https://api.github.com/users/migueltorrescosta/events{/privacy}", "received_events_url": "https://api.github.com/users/migueltorrescosta/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
open
false
{ "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "login": "pdevine", "id": 75239, "node_id": "MDQ6VXNlcjc1MjM5", "avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pdevine", "html_url": "https://github.com/pdevine", "followers_url": "https://api.github.com/users/pdevine/followers", "following_url": "https://api.github.com/users/pdevine/following{/other_user}", "gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}", "starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/pdevine/subscriptions", "organizations_url": "https://api.github.com/users/pdevine/orgs", "repos_url": "https://api.github.com/users/pdevine/repos", "events_url": "https://api.github.com/users/pdevine/events{/privacy}", "received_events_url": "https://api.github.com/users/pdevine/received_events", "type": "User", "user_view_type": "public", "site_admin": false } ]
null
1
2025-01-02T12:30:26
2025-01-08T18:01:09
null
NONE
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
Having ollama models keep track of past conversations improves the quality of answers dramatically. Using `/save <model>` manually with every run is time-consuming, and prone to be forgotten. Allowing a flag similar to `ollama run --auto-save-on-exit <model>` would make this process a lot smoother. The solutions proposed on https://github.com/ollama/ollama/issues/1238 and https://github.com/ollama/ollama/issues/3800 are not sufficient to satisfy this feature request. Thank you for the hard work already done 🙏
null
{ "url": "https://api.github.com/repos/ollama/ollama/issues/8288/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/8288/timeline
null
null
false
https://api.github.com/repos/ollama/ollama/issues/5425
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5425/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5425/comments
https://api.github.com/repos/ollama/ollama/issues/5425/events
https://github.com/ollama/ollama/issues/5425
2,385,008,360
I_kwDOJ0Z1Ps6OKFbo
5,425
Does having the default quant type being Q4_0 (a legacy format) on the model hub still make sense?
{ "login": "sammcj", "id": 862951, "node_id": "MDQ6VXNlcjg2Mjk1MQ==", "avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sammcj", "html_url": "https://github.com/sammcj", "followers_url": "https://api.github.com/users/sammcj/followers", "following_url": "https://api.github.com/users/sammcj/following{/other_user}", "gists_url": "https://api.github.com/users/sammcj/gists{/gist_id}", "starred_url": "https://api.github.com/users/sammcj/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sammcj/subscriptions", "organizations_url": "https://api.github.com/users/sammcj/orgs", "repos_url": "https://api.github.com/users/sammcj/repos", "events_url": "https://api.github.com/users/sammcj/events{/privacy}", "received_events_url": "https://api.github.com/users/sammcj/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[ { "id": 5667396200, "node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA", "url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request", "name": "feature request", "color": "a2eeef", "default": false, "description": "New feature or request" } ]
closed
false
null
[]
null
5
2024-07-02T01:28:02
2024-12-29T19:18:56
2024-12-29T19:18:56
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
null
null
The Ollama model hub still has the default quant type of Q4_0 which is a legacy format that under-performs compared to K-quants (Qn_K, e.g. Q4_K_M, Q6_K, Q5_K_L etc...). - Would it perhaps make sense to change the default quant to Q4_K_M for future models uploaded to the hub? Reference - https://github.com/ggerganov/llama.cpp/wiki/Feature-matrix (Note that the legacy quant types don't even appear on the feature matrix). - https://github.com/ggerganov/llama.cpp/wiki/Tensor-Encoding-Schemes - https://www.reddit.com/r/LocalLLaMA/comments/1ba55rj/overview_of_gguf_quantization_methods/ - https://github.com/ggerganov/llama.cpp/discussions/406#discussioncomment-6176448 - https://github.com/ggerganov/llama.cpp/discussions/2094 - https://huggingface.co/datasets/christopherthompson81/quant_exploration ![image](https://github.com/ollama/ollama/assets/862951/201f324e-d27e-4fd9-b6d4-408699191685) <img width="684" alt="image" src="https://github.com/ollama/ollama/assets/862951/2ab178bb-399e-4fb6-a628-9e60ffb694ab"> ![image](https://github.com/ollama/ollama/assets/862951/5b8b200f-a68e-4178-a3a8-43f81fb7753b) <img width="949" alt="image" src="https://github.com/ollama/ollama/assets/862951/520fcec7-1808-4893-ae02-401fee430e6e"> (Sorry if an issue already exists for this - if it did my search-foo let me down)
{ "login": "jmorganca", "id": 251292, "node_id": "MDQ6VXNlcjI1MTI5Mg==", "avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jmorganca", "html_url": "https://github.com/jmorganca", "followers_url": "https://api.github.com/users/jmorganca/followers", "following_url": "https://api.github.com/users/jmorganca/following{/other_user}", "gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}", "starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions", "organizations_url": "https://api.github.com/users/jmorganca/orgs", "repos_url": "https://api.github.com/users/jmorganca/repos", "events_url": "https://api.github.com/users/jmorganca/events{/privacy}", "received_events_url": "https://api.github.com/users/jmorganca/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5425/reactions", "total_count": 22, "+1": 22, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5425/timeline
null
completed
false
https://api.github.com/repos/ollama/ollama/issues/5653
https://api.github.com/repos/ollama/ollama
https://api.github.com/repos/ollama/ollama/issues/5653/labels{/name}
https://api.github.com/repos/ollama/ollama/issues/5653/comments
https://api.github.com/repos/ollama/ollama/issues/5653/events
https://github.com/ollama/ollama/pull/5653
2,406,177,059
PR_kwDOJ0Z1Ps51QMhb
5,653
template: preprocess message and collect system
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
[]
closed
false
null
[]
null
0
2024-07-12T19:02:17
2024-07-12T19:32:36
2024-07-12T19:32:34
CONTRIBUTOR
{ "total": 0, "completed": 0, "percent_completed": 0 }
null
false
{ "url": "https://api.github.com/repos/ollama/ollama/pulls/5653", "html_url": "https://github.com/ollama/ollama/pull/5653", "diff_url": "https://github.com/ollama/ollama/pull/5653.diff", "patch_url": "https://github.com/ollama/ollama/pull/5653.patch", "merged_at": "2024-07-12T19:32:34" }
null
{ "login": "mxyng", "id": 2372640, "node_id": "MDQ6VXNlcjIzNzI2NDA=", "avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mxyng", "html_url": "https://github.com/mxyng", "followers_url": "https://api.github.com/users/mxyng/followers", "following_url": "https://api.github.com/users/mxyng/following{/other_user}", "gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}", "starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mxyng/subscriptions", "organizations_url": "https://api.github.com/users/mxyng/orgs", "repos_url": "https://api.github.com/users/mxyng/repos", "events_url": "https://api.github.com/users/mxyng/events{/privacy}", "received_events_url": "https://api.github.com/users/mxyng/received_events", "type": "User", "user_view_type": "public", "site_admin": false }
{ "url": "https://api.github.com/repos/ollama/ollama/issues/5653/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/ollama/ollama/issues/5653/timeline
null
null
true