url
stringlengths 51
54
| repository_url
stringclasses 1
value | labels_url
stringlengths 65
68
| comments_url
stringlengths 60
63
| events_url
stringlengths 58
61
| html_url
stringlengths 39
44
| id
int64 1.78B
2.82B
| node_id
stringlengths 18
19
| number
int64 1
8.69k
| title
stringlengths 1
382
| user
dict | labels
listlengths 0
5
| state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
listlengths 0
2
| milestone
null | comments
int64 0
323
| created_at
timestamp[s] | updated_at
timestamp[s] | closed_at
timestamp[s] | author_association
stringclasses 4
values | sub_issues_summary
dict | active_lock_reason
null | draft
bool 2
classes | pull_request
dict | body
stringlengths 2
118k
⌀ | closed_by
dict | reactions
dict | timeline_url
stringlengths 60
63
| performed_via_github_app
null | state_reason
stringclasses 4
values | is_pull_request
bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/3963
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3963/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3963/comments
|
https://api.github.com/repos/ollama/ollama/issues/3963/events
|
https://github.com/ollama/ollama/pull/3963
| 2,266,554,114
|
PR_kwDOJ0Z1Ps5t4mC3
| 3,963
|
windows: init `cmakeTargets` in `init_vars`
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-04-26T21:57:05
| 2024-04-26T22:40:31
| 2024-04-26T22:40:31
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3963",
"html_url": "https://github.com/ollama/ollama/pull/3963",
"diff_url": "https://github.com/ollama/ollama/pull/3963.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3963.patch",
"merged_at": null
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3963/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3963/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/952
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/952/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/952/comments
|
https://api.github.com/repos/ollama/ollama/issues/952/events
|
https://github.com/ollama/ollama/pull/952
| 1,969,294,048
|
PR_kwDOJ0Z1Ps5eK5xF
| 952
|
support raw generation requests
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-10-30T22:01:26
| 2023-11-08T22:05:03
| 2023-11-08T22:05:02
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/952",
"html_url": "https://github.com/ollama/ollama/pull/952",
"diff_url": "https://github.com/ollama/ollama/pull/952.diff",
"patch_url": "https://github.com/ollama/ollama/pull/952.patch",
"merged_at": "2023-11-08T22:05:02"
}
|
- add the optional `raw` generate request parameter to bypass prompt formatting and response context
Add a `raw` parameter to `/generate` requests that allow directly specifying the prompt without the Ollama server applying additional formatting.
```bash
curl -X "POST" -d '{"model":"mistral", "prompt": "[INST] hi [/INST]", "raw": true, "stream": false}' 'http://127.0.0.1:11434/api/generate'
```
Example use case, few-shot prompting:
```python
import requests
def call_generate_endpoint(prompt, model="mistral", raw=True, stream=False):
url = "http://127.0.0.1:11434/api/generate"
formatted_prompt = f"""[INST] This is awesome! [/INST]
Postive
[INST] This is bad! [/INST]
Negative
[INST] I love this movie [/INST]
Positive
[INST] {prompt} [/INST]
"""
payload = {
"model": model,
"prompt": formatted_prompt,
"raw": raw,
"stream": stream
}
response = requests.post(url, json=payload)
return response.json()
resp = call_generate_endpoint("I hate this book")
print(resp.response) # Negative
```
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/952/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/952/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1082
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1082/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1082/comments
|
https://api.github.com/repos/ollama/ollama/issues/1082/events
|
https://github.com/ollama/ollama/pull/1082
| 1,988,676,727
|
PR_kwDOJ0Z1Ps5fMqVc
| 1,082
|
Add support for Multimodel models
|
{
"login": "mattapperson",
"id": 192433,
"node_id": "MDQ6VXNlcjE5MjQzMw==",
"avatar_url": "https://avatars.githubusercontent.com/u/192433?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mattapperson",
"html_url": "https://github.com/mattapperson",
"followers_url": "https://api.github.com/users/mattapperson/followers",
"following_url": "https://api.github.com/users/mattapperson/following{/other_user}",
"gists_url": "https://api.github.com/users/mattapperson/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mattapperson/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mattapperson/subscriptions",
"organizations_url": "https://api.github.com/users/mattapperson/orgs",
"repos_url": "https://api.github.com/users/mattapperson/repos",
"events_url": "https://api.github.com/users/mattapperson/events{/privacy}",
"received_events_url": "https://api.github.com/users/mattapperson/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 12
| 2023-11-11T02:57:28
| 2023-11-22T00:22:46
| 2023-11-22T00:22:46
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1082",
"html_url": "https://github.com/ollama/ollama/pull/1082",
"diff_url": "https://github.com/ollama/ollama/pull/1082.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1082.patch",
"merged_at": null
}
|
Interactive cli usage:
```bash
/set image add <image id int> <path to image file>
Please tell me what text is in this photo [img-<image id int>]
```
For the API I added support for the `image_data` prop with the type of `[]{id: int, data: string(base64)}`for the generate endpoint.
To support this, modelfile now has a `MMPROJ` key that points to the mmproj file path.
Tested with the following modelfile:
```
FROM ./ggml-model-q4_k.gguf
TEMPLATE """
USER:{{ .Prompt }}
ASSISTANT:
"""
MMPROJ ./mmproj-model-f16.gguf
```
And using the following pre-quantatized model - https://huggingface.co/mys/ggml_llava-v1.5-13b.
<img width="1226" alt="Screenshot 2023-11-09 at 11 30 10 PM" src="https://github.com/jmorganca/ollama/assets/192433/ec62e5e7-999b-44f2-9a32-e6b85ca5b404">
<img width="567" alt="screenshot" src="https://github.com/jmorganca/ollama/assets/192433/681c5cd5-9140-43b1-8383-08823cdbb002">
|
{
"login": "mattapperson",
"id": 192433,
"node_id": "MDQ6VXNlcjE5MjQzMw==",
"avatar_url": "https://avatars.githubusercontent.com/u/192433?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mattapperson",
"html_url": "https://github.com/mattapperson",
"followers_url": "https://api.github.com/users/mattapperson/followers",
"following_url": "https://api.github.com/users/mattapperson/following{/other_user}",
"gists_url": "https://api.github.com/users/mattapperson/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mattapperson/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mattapperson/subscriptions",
"organizations_url": "https://api.github.com/users/mattapperson/orgs",
"repos_url": "https://api.github.com/users/mattapperson/repos",
"events_url": "https://api.github.com/users/mattapperson/events{/privacy}",
"received_events_url": "https://api.github.com/users/mattapperson/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1082/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 1,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1082/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8141
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8141/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8141/comments
|
https://api.github.com/repos/ollama/ollama/issues/8141/events
|
https://github.com/ollama/ollama/issues/8141
| 2,745,757,781
|
I_kwDOJ0Z1Ps6jqPBV
| 8,141
|
command r plus model update
|
{
"login": "olumolu",
"id": 162728301,
"node_id": "U_kgDOCbMJbQ",
"avatar_url": "https://avatars.githubusercontent.com/u/162728301?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/olumolu",
"html_url": "https://github.com/olumolu",
"followers_url": "https://api.github.com/users/olumolu/followers",
"following_url": "https://api.github.com/users/olumolu/following{/other_user}",
"gists_url": "https://api.github.com/users/olumolu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/olumolu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/olumolu/subscriptions",
"organizations_url": "https://api.github.com/users/olumolu/orgs",
"repos_url": "https://api.github.com/users/olumolu/repos",
"events_url": "https://api.github.com/users/olumolu/events{/privacy}",
"received_events_url": "https://api.github.com/users/olumolu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-12-17T18:48:33
| 2024-12-23T08:13:29
| 2024-12-23T08:13:29
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
> https://docs.cohere.com/docs/command-r-plus
They already have updated there model can we have the updated new model in olllama thanks
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8141/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8141/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/5864
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5864/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5864/comments
|
https://api.github.com/repos/ollama/ollama/issues/5864/events
|
https://github.com/ollama/ollama/pull/5864
| 2,423,919,879
|
PR_kwDOJ0Z1Ps52Je6A
| 5,864
|
Bump Go patch version
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-07-22T23:17:38
| 2024-07-22T23:34:22
| 2024-07-22T23:34:19
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5864",
"html_url": "https://github.com/ollama/ollama/pull/5864",
"diff_url": "https://github.com/ollama/ollama/pull/5864.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5864.patch",
"merged_at": "2024-07-22T23:34:18"
}
|
Fixes #5774
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5864/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5864/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/321
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/321/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/321/comments
|
https://api.github.com/repos/ollama/ollama/issues/321/events
|
https://github.com/ollama/ollama/pull/321
| 1,846,003,602
|
PR_kwDOJ0Z1Ps5XruCs
| 321
|
length check for parameters
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-08-10T23:09:25
| 2023-08-10T23:23:11
| 2023-08-10T23:23:10
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/321",
"html_url": "https://github.com/ollama/ollama/pull/321",
"diff_url": "https://github.com/ollama/ollama/pull/321.diff",
"patch_url": "https://github.com/ollama/ollama/pull/321.patch",
"merged_at": "2023-08-10T23:23:10"
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/321/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/321/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1392
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1392/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1392/comments
|
https://api.github.com/repos/ollama/ollama/issues/1392/events
|
https://github.com/ollama/ollama/pull/1392
| 2,026,923,082
|
PR_kwDOJ0Z1Ps5hN7DP
| 1,392
|
chat api endpoint
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2023-12-05T18:57:45
| 2023-12-11T07:33:21
| 2023-12-05T19:57:33
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1392",
"html_url": "https://github.com/ollama/ollama/pull/1392",
"diff_url": "https://github.com/ollama/ollama/pull/1392.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1392.patch",
"merged_at": "2023-12-05T19:57:33"
}
|
- add a new `/api/chat` API endpoint that takes an array of `message` objects. This endpoint is an alternative to `/api/generate`.
- deprecate generation context and template, but continue to support them
- rebuild chat content from messages
This changes adds a `/api/chat` endpoint to the API which takes an array of messages. This makes modifying and tracking the history on the fly much simpler. It is an alternative to prompt/response.
`context` will continue to work as expected for now, but at some point in the future we may want to replace it completely with `/api/chat`.
```
curl -X POST http://localhost:11434/api/generate -d '{
"model": "mistral",
"prompt": "hello, how are you?"
}'
OR
### Basic generate request with messages
curl -X POST http://localhost:11434/api/chat -d '{
"model": "mistral",
"messages": [
{
"role": "user",
"content": "why is the sky blue?"
}
]
}'
```
resolves #981
resolves #1203
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1392/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1392/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1940
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1940/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1940/comments
|
https://api.github.com/repos/ollama/ollama/issues/1940/events
|
https://github.com/ollama/ollama/issues/1940
| 2,077,872,259
|
I_kwDOJ0Z1Ps572dCD
| 1,940
|
`CUDA error 100` after detecting GPU libraries on system
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 10
| 2024-01-12T01:13:13
| 2024-02-19T19:50:24
| 2024-02-19T19:50:24
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
It seems that upon detecting an Nvidia card, `ollama` may error with `CUDA error 100`:
```
Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:88: Detecting GPU type
Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:203: Searching for GPU management library libnvidia-ml.so
Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:248: Discovered GPU libraries: [/usr/lib/x86_64-linux-gnu/libnvidia-ml.so.535.104.05 /usr/lib/wsl/lib/libnvidia-ml.so.1]
Jan 11 15:37:50 LR9135SQP ollama[5616]: 2024/01/11 15:37:50 gpu.go:259: Unable to load CUDA management library /usr/lib/x86_64-linux-gnu/libnvidia-ml.so.535.104.05: nvml vram init failure: 9
Jan 11 15:37:51 LR9135SQP ollama[5616]: 2024/01/11 15:37:51 gpu.go:94: Nvidia GPU detected
Jan 11 15:37:51 LR9135SQP ollama[5616]: 2024/01/11 15:37:51 gpu.go:135: CUDA Compute Capability detected: 7.5
```
```
Jan 11 15:55:41 LR9135SQP ollama[5616]: CUDA error 100 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/ggml-cuda.cu:495: no CUDA-capable device is detected
Jan 11 15:55:41 LR9135SQP ollama[5616]: current device: 1881676272
Jan 11 15:55:41 LR9135SQP ollama[5616]: Lazy loading /tmp/ollama958766944/cuda/libext_server.so library
Jan 11 15:55:41 LR9135SQP ollama[5616]: GGML_ASSERT: /go/src/github.com/jmorganca/ollama/llm/llama.cpp/ggml-cuda.cu:495: !"CUDA error"
```
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1940/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1940/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6839
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6839/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6839/comments
|
https://api.github.com/repos/ollama/ollama/issues/6839/events
|
https://github.com/ollama/ollama/issues/6839
| 2,531,251,465
|
I_kwDOJ0Z1Ps6W39UJ
| 6,839
|
ollama request llama3.1 fail.
|
{
"login": "microbitcswcss",
"id": 129246500,
"node_id": "U_kgDOB7QlJA",
"avatar_url": "https://avatars.githubusercontent.com/u/129246500?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/microbitcswcss",
"html_url": "https://github.com/microbitcswcss",
"followers_url": "https://api.github.com/users/microbitcswcss/followers",
"following_url": "https://api.github.com/users/microbitcswcss/following{/other_user}",
"gists_url": "https://api.github.com/users/microbitcswcss/gists{/gist_id}",
"starred_url": "https://api.github.com/users/microbitcswcss/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/microbitcswcss/subscriptions",
"organizations_url": "https://api.github.com/users/microbitcswcss/orgs",
"repos_url": "https://api.github.com/users/microbitcswcss/repos",
"events_url": "https://api.github.com/users/microbitcswcss/events{/privacy}",
"received_events_url": "https://api.github.com/users/microbitcswcss/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 5
| 2024-09-17T14:00:01
| 2024-10-28T07:39:53
| 2024-10-28T07:39:52
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
like title i cant request llama3.1
on windows 10
it can be used one month ago,
cmd.exe:
C:\Users\123>ollama run llama3.1
2024/09/17 21:53:18 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434
2024/09/17 21:53:18 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434
Error: Head "http://C:11434/": dial tcp: lookup C: no such host
C:\Users\123 >ollama --version
2024/09/17 21:59:13 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434
2024/09/17 21:59:13 config.go:45: WARN invalid port, using default port="\\Users\\123\\AppData\\Local\\Programs\\Ollama\\ollama app.exe" default=11434
Warning: could not connect to a running Ollama instance
Warning: client version is 0.3.9
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.9
|
{
"login": "microbitcswcss",
"id": 129246500,
"node_id": "U_kgDOB7QlJA",
"avatar_url": "https://avatars.githubusercontent.com/u/129246500?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/microbitcswcss",
"html_url": "https://github.com/microbitcswcss",
"followers_url": "https://api.github.com/users/microbitcswcss/followers",
"following_url": "https://api.github.com/users/microbitcswcss/following{/other_user}",
"gists_url": "https://api.github.com/users/microbitcswcss/gists{/gist_id}",
"starred_url": "https://api.github.com/users/microbitcswcss/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/microbitcswcss/subscriptions",
"organizations_url": "https://api.github.com/users/microbitcswcss/orgs",
"repos_url": "https://api.github.com/users/microbitcswcss/repos",
"events_url": "https://api.github.com/users/microbitcswcss/events{/privacy}",
"received_events_url": "https://api.github.com/users/microbitcswcss/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6839/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6839/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/220
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/220/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/220/comments
|
https://api.github.com/repos/ollama/ollama/issues/220/events
|
https://github.com/ollama/ollama/issues/220
| 1,822,783,975
|
I_kwDOJ0Z1Ps5spXnn
| 220
|
Sending input with a + crashes
|
{
"login": "benr75",
"id": 661,
"node_id": "MDQ6VXNlcjY2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/661?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/benr75",
"html_url": "https://github.com/benr75",
"followers_url": "https://api.github.com/users/benr75/followers",
"following_url": "https://api.github.com/users/benr75/following{/other_user}",
"gists_url": "https://api.github.com/users/benr75/gists{/gist_id}",
"starred_url": "https://api.github.com/users/benr75/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/benr75/subscriptions",
"organizations_url": "https://api.github.com/users/benr75/orgs",
"repos_url": "https://api.github.com/users/benr75/repos",
"events_url": "https://api.github.com/users/benr75/events{/privacy}",
"received_events_url": "https://api.github.com/users/benr75/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 4
| 2023-07-26T16:37:12
| 2023-08-04T14:28:22
| 2023-08-04T14:28:22
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Maybe an url escaping issue? I didn't dig much deeper but wanted to report.
```
$ ollama run llama2:13b
>>> I cannot send input with a + in it
Error: Post "http://127.0.0.1:11434/api/generate": EOF
```
|
{
"login": "benr75",
"id": 661,
"node_id": "MDQ6VXNlcjY2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/661?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/benr75",
"html_url": "https://github.com/benr75",
"followers_url": "https://api.github.com/users/benr75/followers",
"following_url": "https://api.github.com/users/benr75/following{/other_user}",
"gists_url": "https://api.github.com/users/benr75/gists{/gist_id}",
"starred_url": "https://api.github.com/users/benr75/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/benr75/subscriptions",
"organizations_url": "https://api.github.com/users/benr75/orgs",
"repos_url": "https://api.github.com/users/benr75/repos",
"events_url": "https://api.github.com/users/benr75/events{/privacy}",
"received_events_url": "https://api.github.com/users/benr75/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/220/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/220/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8248
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8248/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8248/comments
|
https://api.github.com/repos/ollama/ollama/issues/8248/events
|
https://github.com/ollama/ollama/issues/8248
| 2,759,846,975
|
I_kwDOJ0Z1Ps6kf-w_
| 8,248
|
Unexpected Connection Closure and GPU Memory Not Releasing
|
{
"login": "BronyaKaslana06",
"id": 97924591,
"node_id": "U_kgDOBdY17w",
"avatar_url": "https://avatars.githubusercontent.com/u/97924591?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BronyaKaslana06",
"html_url": "https://github.com/BronyaKaslana06",
"followers_url": "https://api.github.com/users/BronyaKaslana06/followers",
"following_url": "https://api.github.com/users/BronyaKaslana06/following{/other_user}",
"gists_url": "https://api.github.com/users/BronyaKaslana06/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BronyaKaslana06/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BronyaKaslana06/subscriptions",
"organizations_url": "https://api.github.com/users/BronyaKaslana06/orgs",
"repos_url": "https://api.github.com/users/BronyaKaslana06/repos",
"events_url": "https://api.github.com/users/BronyaKaslana06/events{/privacy}",
"received_events_url": "https://api.github.com/users/BronyaKaslana06/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 7
| 2024-12-26T14:40:51
| 2025-01-01T02:27:55
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
### Problem Description
I am using Ubuntu 22.04 and making network requests to a local Ollama service with Python to run a series of models sequentially. After each model runs, it is unloaded using the following Python code. However, the task encounters an issue after reaching a certain point, where the Ollama network service unexpectedly stops when loading a model, resulting in the following error:
```
Traceback (most recent call last):
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 703, in urlopen
httplib_response = self._make_request(
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 449, in _make_request
six.raise_from(e, None)
File "<string>", line 3, in raise_from
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 444, in _make_request
httplib_response = conn.getresponse()
File "/root/miniconda3/lib/python3.10/http/client.py", line 1374, in getresponse
response.begin()
File "/root/miniconda3/lib/python3.10/http/client.py", line 318, in begin
version, status, reason = self._read_status()
File "/root/miniconda3/lib/python3.10/http/client.py", line 287, in _read_status
raise RemoteDisconnected("Remote end closed connection without"
http.client.RemoteDisconnected: Remote end closed connection without response
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/root/miniconda3/lib/python3.10/site-packages/requests/adapters.py", line 667, in send
resp = conn.urlopen(
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 787, in urlopen
retries = retries.increment(
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/util/retry.py", line 550, in increment
raise six.reraise(type(error), error, _stacktrace)
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/packages/six.py", line 769, in reraise
raise value.with_traceback(tb)
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 703, in urlopen
httplib_response = self._make_request(
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 449, in _make_request
six.raise_from(e, None)
File "<string>", line 3, in raise_from
File "/root/miniconda3/lib/python3.10/site-packages/urllib3/connectionpool.py", line 444, in _make_request
httplib_response = conn.getresponse()
File "/root/miniconda3/lib/python3.10/http/client.py", line 1374, in getresponse
response.begin()
File "/root/miniconda3/lib/python3.10/http/client.py", line 318, in begin
version, status, reason = self._read_status()
File "/root/miniconda3/lib/python3.10/http/client.py", line 287, in _read_status
raise RemoteDisconnected("Remote end closed connection without"
urllib3.exceptions.ProtocolError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/root/autodl-tmp/EvalLLM/main.py", line 109, in <module>
generator.generate_and_save(section)
File "/root/autodl-tmp/EvalLLM/main.py", line 93, in generate_and_save
response_text = ollamaCaller.generate_response(prompt=question)
File "/root/autodl-tmp/EvalLLM/Caller/OllamaCaller.py", line 18, in generate_response
response = requests.post("http://localhost:11434/api/generate", json=payload)
File "/root/miniconda3/lib/python3.10/site-packages/requests/api.py", line 115, in post
return request("post", url, data=data, json=json, **kwargs)
File "/root/miniconda3/lib/python3.10/site-packages/requests/api.py", line 59, in request
return session.request(method=method, url=url, **kwargs)
File "/root/miniconda3/lib/python3.10/site-packages/requests/sessions.py", line 589, in request
resp = self.send(prep, **send_kwargs)
File "/root/miniconda3/lib/python3.10/site-packages/requests/sessions.py", line 703, in send
r = adapter.send(request, **kwargs)
File "/root/miniconda3/lib/python3.10/site-packages/requests/adapters.py", line 682, in send
raise ConnectionError(err, request=request)
requests.exceptions.ConnectionError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response'))
```
This could be caused by the **service not responding for a long time**, or it might be due to the **daemon process being unexpectedly closed**, right? At the time of the error, The program was just finished loading a new model (**a relatively small model for my gpu, llama3:latest**), and the Ollama logs are as follows:
```log
time=2024-12-25T22:05:34.725+08:00 level=INFO source=sched.go:714 msg="new model will fit in available VRAM in single GPU, loading" model=/root/autodl-tmp/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa gpu=GPU-30d02008-3980-d577-cd07-7e3fcabf1d31 parallel=4 available=25158156288 required="6.2 GiB"
time=2024-12-25T22:05:34.977+08:00 level=INFO source=server.go:104 msg="system memory" total="377.3 GiB" free="327.5 GiB" free_swap="0 B"
time=2024-12-25T22:05:34.977+08:00 level=INFO source=memory.go:356 msg="offload to cuda" layers.requested=-1 layers.model=33 layers.offload=33 layers.split="" memory.available="[23.4 GiB]" memory.gpu_overhead="0 B" memory.required.full="6.2 GiB" memory.required.partial="6.2 GiB" memory.required.kv="1.0 GiB" memory.required.allocations="[6.2 GiB]" memory.weights.total="4.7 GiB" memory.weights.repeating="4.3 GiB" memory.weights.nonrepeating="411.0 MiB" memory.graph.full="560.0 MiB" memory.graph.partial="677.5 MiB"
time=2024-12-25T22:05:34.977+08:00 level=INFO source=server.go:376 msg="starting llama server" cmd="/usr/local/lib/ollama/runners/cuda_v12_avx/ollama_llama_server runner --model /root/autodl-tmp/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa --ctx-size 8192 --batch-size 512 --n-gpu-layers 33 --threads 112 --parallel 4 --port 41461"
time=2024-12-25T22:05:34.978+08:00 level=INFO source=sched.go:449 msg="loaded runners" count=1
time=2024-12-25T22:05:34.978+08:00 level=INFO source=server.go:555 msg="waiting for llama runner to start responding"
time=2024-12-25T22:05:34.978+08:00 level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server error"
time=2024-12-25T22:05:35.021+08:00 level=INFO source=runner.go:945 msg="starting go runner"
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes
time=2024-12-25T22:05:35.029+08:00 level=INFO source=runner.go:946 msg=system info="CUDA : ARCHS = 600,610,620,700,720,750,800,860,870,890,900 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | LLAMAFILE = 1 | AARCH64_REPACK = 1 | cgo(gcc)" threads=112
time=2024-12-25T22:05:35.029+08:00 level=INFO source=.:0 msg="Server listening on 127.0.0.1:41461"
llama_load_model_from_file: using device CUDA0 (NVIDIA GeForce RTX 3090) - 23992 MiB free
llama_model_loader: loaded meta data with 22 key-value pairs and 291 tensors from /root/autodl-tmp/blobs/sha256-6a0746a1ec1aef3e7ec53868f220ff6e389f6f8ef87a01d77c96807de94ca2aa (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = llama
llama_model_loader: - kv 1: general.name str = Meta-Llama-3-8B-Instruct
llama_model_loader: - kv 2: llama.block_count u32 = 32
llama_model_loader: - kv 3: llama.context_length u32 = 8192
llama_model_loader: - kv 4: llama.embedding_length u32 = 4096
llama_model_loader: - kv 5: llama.feed_forward_length u32 = 14336
llama_model_loader: - kv 6: llama.attention.head_count u32 = 32
llama_model_loader: - kv 7: llama.attention.head_count_kv u32 = 8
llama_model_loader: - kv 8: llama.rope.freq_base f32 = 500000.000000
llama_model_loader: - kv 9: llama.attention.layer_norm_rms_epsilon f32 = 0.000010
llama_model_loader: - kv 10: general.file_type u32 = 2
llama_model_loader: - kv 11: llama.vocab_size u32 = 128256
llama_model_loader: - kv 12: llama.rope.dimension_count u32 = 128
llama_model_loader: - kv 13: tokenizer.ggml.model str = gpt2
llama_model_loader: - kv 14: tokenizer.ggml.pre str = llama-bpe
llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,128256] = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv 16: tokenizer.ggml.token_type arr[i32,128256] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
time=2024-12-25T22:05:35.230+08:00 level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server loading model"
llama_model_loader: - kv 17: tokenizer.ggml.merges arr[str,280147] = ["Ġ Ġ", "Ġ ĠĠĠ", "ĠĠ ĠĠ", "...
llama_model_loader: - kv 18: tokenizer.ggml.bos_token_id u32 = 128000
llama_model_loader: - kv 19: tokenizer.ggml.eos_token_id u32 = 128009
llama_model_loader: - kv 20: tokenizer.chat_template str = {% set loop_messages = messages %}{% ...
llama_model_loader: - kv 21: general.quantization_version u32 = 2
llama_model_loader: - type f32: 65 tensors
llama_model_loader: - type q4_0: 225 tensors
llama_model_loader: - type q6_K: 1 tensors
llm_load_vocab: special tokens cache size = 256
llm_load_vocab: token to piece cache size = 0.8000 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = llama
llm_load_print_meta: vocab type = BPE
llm_load_print_meta: n_vocab = 128256
llm_load_print_meta: n_merges = 280147
llm_load_print_meta: vocab_only = 0
llm_load_print_meta: n_ctx_train = 8192
llm_load_print_meta: n_embd = 4096
llm_load_print_meta: n_layer = 32
llm_load_print_meta: n_head = 32
llm_load_print_meta: n_head_kv = 8
llm_load_print_meta: n_rot = 128
llm_load_print_meta: n_swa = 0
llm_load_print_meta: n_embd_head_k = 128
llm_load_print_meta: n_embd_head_v = 128
llm_load_print_meta: n_gqa = 4
llm_load_print_meta: n_embd_k_gqa = 1024
llm_load_print_meta: n_embd_v_gqa = 1024
llm_load_print_meta: f_norm_eps = 0.0e+00
llm_load_print_meta: f_norm_rms_eps = 1.0e-05
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 14336
llm_load_print_meta: n_expert = 0
llm_load_print_meta: n_expert_used = 0
llm_load_print_meta: causal attn = 1
llm_load_print_meta: pooling type = 0
llm_load_print_meta: rope type = 0
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 500000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_ctx_orig_yarn = 8192
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: ssm_dt_b_c_rms = 0
llm_load_print_meta: model type = 8B
llm_load_print_meta: model ftype = Q4_0
llm_load_print_meta: model params = 8.03 B
llm_load_print_meta: model size = 4.33 GiB (4.64 BPW)
llm_load_print_meta: general.name = Meta-Llama-3-8B-Instruct
llm_load_print_meta: BOS token = 128000 '<|begin_of_text|>'
llm_load_print_meta: EOS token = 128009 '<|eot_id|>'
llm_load_print_meta: EOT token = 128009 '<|eot_id|>'
llm_load_print_meta: LF token = 128 'Ä'
llm_load_print_meta: EOG token = 128009 '<|eot_id|>'
llm_load_print_meta: max token length = 256
llm_load_tensors: offloading 32 repeating layers to GPU
llm_load_tensors: offloading output layer to GPU
llm_load_tensors: offloaded 33/33 layers to GPU
llm_load_tensors: CPU_Mapped model buffer size = 281.81 MiB
llm_load_tensors: CUDA0 model buffer size = 4155.99 MiB
llama_new_context_with_model: n_seq_max = 4
llama_new_context_with_model: n_ctx = 8192
llama_new_context_with_model: n_ctx_per_seq = 2048
llama_new_context_with_model: n_batch = 2048
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 500000.0
llama_new_context_with_model: freq_scale = 1
llama_new_context_with_model: n_ctx_per_seq (2048) < n_ctx_train (8192) -- the full capacity of the model will not be utilized
llama_kv_cache_init: CUDA0 KV buffer size = 1024.00 MiB
llama_new_context_with_model: KV self size = 1024.00 MiB, K (f16): 512.00 MiB, V (f16): 512.00 MiB
llama_new_context_with_model: CUDA_Host output buffer size = 2.02 MiB
llama_new_context_with_model: CUDA0 compute buffer size = 560.00 MiB
llama_new_context_with_model: CUDA_Host compute buffer size = 24.01 MiB
llama_new_context_with_model: graph nodes = 1030
llama_new_context_with_model: graph splits = 2
```
The logs end here.
Regardless of the cause of this issue, after it occurs, **about 6GB of GPU memory remains unreleased and cannot be freed even by restarting Ollama**. Below is the output of `nvidia-smi`:
```
(base) root@autodl-container-d33848b29e-31b2d2f4:~# nvidia-smi
Thu Dec 26 22:29:15 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.78 Driver Version: 550.78 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 3090 On | 00000000:10:00.0 Off | N/A |
| 30% 28C P8 17W / 350W | 6002MiB / 24576MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
+-----------------------------------------------------------------------------------------+
```
Has anyone encountered this issue? How can it be resolved? Thank you!
### OS
Linux
### GPU
Nvidia
### CPU
_No response_
### Ollama version
0.5.4
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8248/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8248/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/7455
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7455/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7455/comments
|
https://api.github.com/repos/ollama/ollama/issues/7455/events
|
https://github.com/ollama/ollama/pull/7455
| 2,627,779,798
|
PR_kwDOJ0Z1Ps6AkeyU
| 7,455
|
Add basic mllama integration tests
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-10-31T20:37:06
| 2024-11-01T00:25:51
| 2024-11-01T00:25:48
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7455",
"html_url": "https://github.com/ollama/ollama/pull/7455",
"diff_url": "https://github.com/ollama/ollama/pull/7455.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7455.patch",
"merged_at": "2024-11-01T00:25:48"
}
|
Draft until I run it against a few more systems and make sure the timeout values are reasonable
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7455/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7455/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8379
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8379/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8379/comments
|
https://api.github.com/repos/ollama/ollama/issues/8379/events
|
https://github.com/ollama/ollama/issues/8379
| 2,781,273,717
|
I_kwDOJ0Z1Ps6lxt51
| 8,379
|
Inquiry About Adding vikhyatk/moondream2 to Ollama
|
{
"login": "Saber120",
"id": 108297159,
"node_id": "U_kgDOBnR7xw",
"avatar_url": "https://avatars.githubusercontent.com/u/108297159?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Saber120",
"html_url": "https://github.com/Saber120",
"followers_url": "https://api.github.com/users/Saber120/followers",
"following_url": "https://api.github.com/users/Saber120/following{/other_user}",
"gists_url": "https://api.github.com/users/Saber120/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Saber120/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Saber120/subscriptions",
"organizations_url": "https://api.github.com/users/Saber120/orgs",
"repos_url": "https://api.github.com/users/Saber120/repos",
"events_url": "https://api.github.com/users/Saber120/events{/privacy}",
"received_events_url": "https://api.github.com/users/Saber120/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 2
| 2025-01-10T22:03:19
| 2025-01-13T19:29:39
| 2025-01-13T19:29:38
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi Ollama team, I was wondering if you plan to add the model vikhyatk/moondream2, which was just released yesterday, to your platform
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8379/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8379/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1753
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1753/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1753/comments
|
https://api.github.com/repos/ollama/ollama/issues/1753/events
|
https://github.com/ollama/ollama/issues/1753
| 2,061,198,535
|
I_kwDOJ0Z1Ps5622TH
| 1,753
|
[enhancement] Support FROM gpt-4:0314
|
{
"login": "nathanleclaire",
"id": 1476820,
"node_id": "MDQ6VXNlcjE0NzY4MjA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1476820?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nathanleclaire",
"html_url": "https://github.com/nathanleclaire",
"followers_url": "https://api.github.com/users/nathanleclaire/followers",
"following_url": "https://api.github.com/users/nathanleclaire/following{/other_user}",
"gists_url": "https://api.github.com/users/nathanleclaire/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nathanleclaire/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nathanleclaire/subscriptions",
"organizations_url": "https://api.github.com/users/nathanleclaire/orgs",
"repos_url": "https://api.github.com/users/nathanleclaire/repos",
"events_url": "https://api.github.com/users/nathanleclaire/events{/privacy}",
"received_events_url": "https://api.github.com/users/nathanleclaire/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-01-01T00:11:55
| 2024-01-02T11:26:07
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
It would be nice if you could have a `Modelfile` that had an OpenAI model as a base. Then Ollama inference would call the OpenAI APIs instead of local inference with the built parameters. I know it seems like maybe it sorta defeats the point but I really like having the "Ollama facade" available. I use the March version of GPT-4 heavily but have to type "Let's think step by step" and all those other things all the time 😵💫
I could look at making a PR for this if there is interest. @jmorganca @mchiang0610
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1753/reactions",
"total_count": 5,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 5,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1753/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/226
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/226/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/226/comments
|
https://api.github.com/repos/ollama/ollama/issues/226/events
|
https://github.com/ollama/ollama/pull/226
| 1,824,955,065
|
PR_kwDOJ0Z1Ps5Wk6Px
| 226
|
refactor scan multiline for reuse
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-27T18:32:03
| 2023-07-27T18:45:45
| 2023-07-27T18:45:41
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/226",
"html_url": "https://github.com/ollama/ollama/pull/226",
"diff_url": "https://github.com/ollama/ollama/pull/226.diff",
"patch_url": "https://github.com/ollama/ollama/pull/226.patch",
"merged_at": "2023-07-27T18:45:41"
}
|
It's not obvious values are ingested verbatim when not using multiline so `"` are included in the template. Instead, ingest the value inside the quotes
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/226/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/226/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1866
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1866/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1866/comments
|
https://api.github.com/repos/ollama/ollama/issues/1866/events
|
https://github.com/ollama/ollama/issues/1866
| 2,072,359,110
|
I_kwDOJ0Z1Ps57hbDG
| 1,866
|
Where is the model file stored?
|
{
"login": "ParisNeo",
"id": 827993,
"node_id": "MDQ6VXNlcjgyNzk5Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/827993?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParisNeo",
"html_url": "https://github.com/ParisNeo",
"followers_url": "https://api.github.com/users/ParisNeo/followers",
"following_url": "https://api.github.com/users/ParisNeo/following{/other_user}",
"gists_url": "https://api.github.com/users/ParisNeo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParisNeo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParisNeo/subscriptions",
"organizations_url": "https://api.github.com/users/ParisNeo/orgs",
"repos_url": "https://api.github.com/users/ParisNeo/repos",
"events_url": "https://api.github.com/users/ParisNeo/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParisNeo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 4
| 2024-01-09T13:19:37
| 2024-01-11T16:49:16
| 2024-01-09T17:30:51
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi there, I need to make a modification to the model file. Can you please tell me where do you store the model file?
Best regards
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1866/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1866/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5531
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5531/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5531/comments
|
https://api.github.com/repos/ollama/ollama/issues/5531/events
|
https://github.com/ollama/ollama/issues/5531
| 2,394,121,820
|
I_kwDOJ0Z1Ps6Os2Zc
| 5,531
|
`ollama create --quantize` does not show proper error if quantizing an unsupported model architecture
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/joshyan1/followers",
"following_url": "https://api.github.com/users/joshyan1/following{/other_user}",
"gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions",
"organizations_url": "https://api.github.com/users/joshyan1/orgs",
"repos_url": "https://api.github.com/users/joshyan1/repos",
"events_url": "https://api.github.com/users/joshyan1/events{/privacy}",
"received_events_url": "https://api.github.com/users/joshyan1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/joshyan1/followers",
"following_url": "https://api.github.com/users/joshyan1/following{/other_user}",
"gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions",
"organizations_url": "https://api.github.com/users/joshyan1/orgs",
"repos_url": "https://api.github.com/users/joshyan1/repos",
"events_url": "https://api.github.com/users/joshyan1/events{/privacy}",
"received_events_url": "https://api.github.com/users/joshyan1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 0
| 2024-07-07T16:27:14
| 2024-07-12T00:24:30
| 2024-07-12T00:24:30
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
When quantizing a model with `ollama create -q`, unknown model architecture errors are not properly bubbled up. E.g. from the logs:
```
llama_model_loader: - type f32: 121 tensors
llama_model_loader: - type f16: 162 tensors
llama_model_quantize: failed to quantize: unknown model architecture: 'chatglm'
```
User error:
```
% ollama create -f Modelfile -q q4_0 test
transferring model data
quantizing F16 model to Q4_0
Error: llama_model_quantize: 1
```
### OS
_No response_
### GPU
_No response_
### CPU
_No response_
### Ollama version
_No response_
|
{
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/joshyan1/followers",
"following_url": "https://api.github.com/users/joshyan1/following{/other_user}",
"gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions",
"organizations_url": "https://api.github.com/users/joshyan1/orgs",
"repos_url": "https://api.github.com/users/joshyan1/repos",
"events_url": "https://api.github.com/users/joshyan1/events{/privacy}",
"received_events_url": "https://api.github.com/users/joshyan1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5531/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5531/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6806
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6806/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6806/comments
|
https://api.github.com/repos/ollama/ollama/issues/6806/events
|
https://github.com/ollama/ollama/issues/6806
| 2,526,589,141
|
I_kwDOJ0Z1Ps6WmLDV
| 6,806
|
slow
|
{
"login": "ayttop",
"id": 178673810,
"node_id": "U_kgDOCqZYkg",
"avatar_url": "https://avatars.githubusercontent.com/u/178673810?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ayttop",
"html_url": "https://github.com/ayttop",
"followers_url": "https://api.github.com/users/ayttop/followers",
"following_url": "https://api.github.com/users/ayttop/following{/other_user}",
"gists_url": "https://api.github.com/users/ayttop/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ayttop/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ayttop/subscriptions",
"organizations_url": "https://api.github.com/users/ayttop/orgs",
"repos_url": "https://api.github.com/users/ayttop/repos",
"events_url": "https://api.github.com/users/ayttop/events{/privacy}",
"received_events_url": "https://api.github.com/users/ayttop/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] |
closed
| false
| null |
[] | null | 9
| 2024-09-14T20:03:00
| 2024-09-25T21:08:43
| 2024-09-25T21:08:43
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
The program is very slow
3.10 ollama
### OS
Windows
### GPU
Intel
### CPU
Intel
### Ollama version
3.10
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6806/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6806/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1471
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1471/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1471/comments
|
https://api.github.com/repos/ollama/ollama/issues/1471/events
|
https://github.com/ollama/ollama/issues/1471
| 2,036,377,833
|
I_kwDOJ0Z1Ps55YKjp
| 1,471
|
Publishing model fails with "Error: unable to push ..."
|
{
"login": "stephenwithav",
"id": 54563,
"node_id": "MDQ6VXNlcjU0NTYz",
"avatar_url": "https://avatars.githubusercontent.com/u/54563?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stephenwithav",
"html_url": "https://github.com/stephenwithav",
"followers_url": "https://api.github.com/users/stephenwithav/followers",
"following_url": "https://api.github.com/users/stephenwithav/following{/other_user}",
"gists_url": "https://api.github.com/users/stephenwithav/gists{/gist_id}",
"starred_url": "https://api.github.com/users/stephenwithav/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stephenwithav/subscriptions",
"organizations_url": "https://api.github.com/users/stephenwithav/orgs",
"repos_url": "https://api.github.com/users/stephenwithav/repos",
"events_url": "https://api.github.com/users/stephenwithav/events{/privacy}",
"received_events_url": "https://api.github.com/users/stephenwithav/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 5
| 2023-12-11T19:25:53
| 2024-11-03T01:37:33
| 2024-01-20T00:25:23
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I followed the [import doc](https://github.com/jmorganca/ollama/blob/7a1b37ac64f0fb0585e279a0a840707843511ed3/docs/import.md?plain=1#L108) steps, but I'm getting the error in the title.
I've:
1. Created an account.
2. Copied my `id_ed25519.pub` file to `~/.ollama/` and `/usr/share/ollama/.ollama`.
3. Added the public key to my account.
4. Copied the model to my username's namespace.
5. Attempted `ollama push stephenwithav/book-summary`
```
Error: unable to push stephenwithav/book-summary, make sure this namespace exists and you are authorized to push to it
```
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1471/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1471/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/3923
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3923/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3923/comments
|
https://api.github.com/repos/ollama/ollama/issues/3923/events
|
https://github.com/ollama/ollama/pull/3923
| 2,264,567,354
|
PR_kwDOJ0Z1Ps5tx0Ne
| 3,923
|
precalculate output tensor memory for metal and mmap
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-04-25T22:00:38
| 2024-04-25T23:34:18
| 2024-04-25T23:34:17
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3923",
"html_url": "https://github.com/ollama/ollama/pull/3923",
"diff_url": "https://github.com/ollama/ollama/pull/3923.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3923.patch",
"merged_at": "2024-04-25T23:34:17"
}
|
on metal with mmap, the output tensors are always allocated even if the offloaded layers < total layers + 1. other backends are unaffected
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3923/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3923/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4428
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4428/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4428/comments
|
https://api.github.com/repos/ollama/ollama/issues/4428/events
|
https://github.com/ollama/ollama/issues/4428
| 2,295,167,222
|
I_kwDOJ0Z1Ps6IzXj2
| 4,428
|
Ollama Api not working after adding format:"json"
|
{
"login": "usmandilmeer",
"id": 51738693,
"node_id": "MDQ6VXNlcjUxNzM4Njkz",
"avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/usmandilmeer",
"html_url": "https://github.com/usmandilmeer",
"followers_url": "https://api.github.com/users/usmandilmeer/followers",
"following_url": "https://api.github.com/users/usmandilmeer/following{/other_user}",
"gists_url": "https://api.github.com/users/usmandilmeer/gists{/gist_id}",
"starred_url": "https://api.github.com/users/usmandilmeer/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/usmandilmeer/subscriptions",
"organizations_url": "https://api.github.com/users/usmandilmeer/orgs",
"repos_url": "https://api.github.com/users/usmandilmeer/repos",
"events_url": "https://api.github.com/users/usmandilmeer/events{/privacy}",
"received_events_url": "https://api.github.com/users/usmandilmeer/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-05-14T11:23:41
| 2024-05-14T21:46:32
| 2024-05-14T21:46:32
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Ollama api is giving empty response if i add add format as json

### OS
Windows
### GPU
AMD
### CPU
Intel
### Ollama version
0.1.37
|
{
"login": "usmandilmeer",
"id": 51738693,
"node_id": "MDQ6VXNlcjUxNzM4Njkz",
"avatar_url": "https://avatars.githubusercontent.com/u/51738693?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/usmandilmeer",
"html_url": "https://github.com/usmandilmeer",
"followers_url": "https://api.github.com/users/usmandilmeer/followers",
"following_url": "https://api.github.com/users/usmandilmeer/following{/other_user}",
"gists_url": "https://api.github.com/users/usmandilmeer/gists{/gist_id}",
"starred_url": "https://api.github.com/users/usmandilmeer/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/usmandilmeer/subscriptions",
"organizations_url": "https://api.github.com/users/usmandilmeer/orgs",
"repos_url": "https://api.github.com/users/usmandilmeer/repos",
"events_url": "https://api.github.com/users/usmandilmeer/events{/privacy}",
"received_events_url": "https://api.github.com/users/usmandilmeer/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4428/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4428/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5588
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5588/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5588/comments
|
https://api.github.com/repos/ollama/ollama/issues/5588/events
|
https://github.com/ollama/ollama/pull/5588
| 2,399,649,431
|
PR_kwDOJ0Z1Ps506G4S
| 5,588
|
Remove `GGML_CUDA_FORCE_MMQ=on` from build
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-07-10T03:08:09
| 2024-07-10T20:17:14
| 2024-07-10T20:17:13
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5588",
"html_url": "https://github.com/ollama/ollama/pull/5588",
"diff_url": "https://github.com/ollama/ollama/pull/5588.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5588.patch",
"merged_at": "2024-07-10T20:17:13"
}
|
Removing this build flag seems to cause mmq to be decided at runtime vs always forcing it on. This stops some cuda calls from failing on V100 and other CC 7.0 cards. This may also help with [RDNA3 cards](https://github.com/ggerganov/llama.cpp/blob/a59f8fdc85e1119d470d8766e29617962549d993/docs/build.md?plain=1#L175)
Fixes https://github.com/ollama/ollama/issues/5571
Changes to mmq functionality related to this bug: https://github.com/ggerganov/llama.cpp/compare/7c26775...a8db2a9c#diff-b2fe862fcd5119199ae59ea13d1b6a46e0d23e41e727e39d90913f828a5ff66bR78
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5588/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5588/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1614
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1614/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1614/comments
|
https://api.github.com/repos/ollama/ollama/issues/1614/events
|
https://github.com/ollama/ollama/pull/1614
| 2,049,321,465
|
PR_kwDOJ0Z1Ps5iZ9OH
| 1,614
|
fix: set template without triple quotes
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2023-12-19T19:37:53
| 2024-01-09T17:36:26
| 2024-01-09T17:36:25
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1614",
"html_url": "https://github.com/ollama/ollama/pull/1614",
"diff_url": "https://github.com/ollama/ollama/pull/1614.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1614.patch",
"merged_at": "2024-01-09T17:36:25"
}
|
this changes updates `/set` to better handle multiline strings. `/set` now correctly sets template or system without using triple quotes
```
>>> /set template {{ .Prompt }}
Set prompt template.
```
additionally, use a strings.Builder instead of concatenating string values for prompt building
```
>>> """hello
... world"""
```
```
>>> """
... hello
... world
... """
```
```
>>> /set system """
... you are a llama
... """
```
```
>>> /set template """
... {{.System}}
... User: {{.Prompt}}
... Assistant: {{.Response}}
... """
```
resolves #1609
resolves #1607
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1614/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1614/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/473
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/473/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/473/comments
|
https://api.github.com/repos/ollama/ollama/issues/473/events
|
https://github.com/ollama/ollama/pull/473
| 1,882,917,488
|
PR_kwDOJ0Z1Ps5Zn-ay
| 473
|
create manifests directory
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-09-06T00:12:07
| 2023-09-06T00:37:42
| 2023-09-06T00:37:41
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/473",
"html_url": "https://github.com/ollama/ollama/pull/473",
"diff_url": "https://github.com/ollama/ollama/pull/473.diff",
"patch_url": "https://github.com/ollama/ollama/pull/473.patch",
"merged_at": "2023-09-06T00:37:41"
}
|
`ollama list` on a brand new install will panic because the manifests directory doesn't exist
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/473/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/473/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1070
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1070/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1070/comments
|
https://api.github.com/repos/ollama/ollama/issues/1070/events
|
https://github.com/ollama/ollama/issues/1070
| 1,986,849,875
|
I_kwDOJ0Z1Ps52bOxT
| 1,070
|
API for models on `ollama.com`
|
{
"login": "danemadsen",
"id": 11537699,
"node_id": "MDQ6VXNlcjExNTM3Njk5",
"avatar_url": "https://avatars.githubusercontent.com/u/11537699?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/danemadsen",
"html_url": "https://github.com/danemadsen",
"followers_url": "https://api.github.com/users/danemadsen/followers",
"following_url": "https://api.github.com/users/danemadsen/following{/other_user}",
"gists_url": "https://api.github.com/users/danemadsen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/danemadsen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/danemadsen/subscriptions",
"organizations_url": "https://api.github.com/users/danemadsen/orgs",
"repos_url": "https://api.github.com/users/danemadsen/repos",
"events_url": "https://api.github.com/users/danemadsen/events{/privacy}",
"received_events_url": "https://api.github.com/users/danemadsen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] |
closed
| false
| null |
[] | null | 12
| 2023-11-10T04:23:05
| 2024-11-21T17:25:46
| 2024-11-21T17:25:46
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi I was wondering if you could add a way to either search for, or get a list of models available to pull off ollama.ai.
Currently the https://ollama.ai/library endpoint serves model information as HTML, it would be better if it was served as JSON so it could be better integrated into software
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1070/reactions",
"total_count": 18,
"+1": 18,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1070/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6120
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6120/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6120/comments
|
https://api.github.com/repos/ollama/ollama/issues/6120/events
|
https://github.com/ollama/ollama/issues/6120
| 2,442,618,478
|
I_kwDOJ0Z1Ps6Rl2Zu
| 6,120
|
Outputting the response leaves a bunch of control characters.
|
{
"login": "jjasghar",
"id": 810824,
"node_id": "MDQ6VXNlcjgxMDgyNA==",
"avatar_url": "https://avatars.githubusercontent.com/u/810824?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jjasghar",
"html_url": "https://github.com/jjasghar",
"followers_url": "https://api.github.com/users/jjasghar/followers",
"following_url": "https://api.github.com/users/jjasghar/following{/other_user}",
"gists_url": "https://api.github.com/users/jjasghar/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jjasghar/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jjasghar/subscriptions",
"organizations_url": "https://api.github.com/users/jjasghar/orgs",
"repos_url": "https://api.github.com/users/jjasghar/repos",
"events_url": "https://api.github.com/users/jjasghar/events{/privacy}",
"received_events_url": "https://api.github.com/users/jjasghar/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 9
| 2024-08-01T14:25:11
| 2024-11-22T16:04:55
| 2024-11-22T16:04:55
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I get a bunch of control characters. How can I get a clean output?
_Originally posted by @pongnguy in https://github.com/ollama/ollama/issues/1710#issuecomment-2142294830_
It seems when you run something like: `ollama run granite-code:20b show me a python fuction that does fizzbuzz > out.txt` you end up with a bunch of:
```
^[[?25l^[[?25hThis^[[?25l^[[?25h function^[[?25l^[[?25h takes^[[?2
5l^[[?25h a^[[?25l^[[?25h positive^[[?25l^[[?25h integer^[[?25l^[[
?25h `^[[?25l^[[?25hn^[[?25l^[[?25h`^[[?25l^[[?25h as^[[?25l^[[?25
h input^[[?25l^[[?25h and^[[?25l^[[?25h prints^[[?25l^[[?25h the^[
[?25l^[[?25h numbers^[[?25l^[[?25h from^[[?25l^[[?25h ^[[?25l^[[?2
5h1^[[?25l^[[?25h to^[[?25l^[[?25h `^[[?25l^[[?25hn^[[?25l^[[?25h`
,^[[?25l^[[?25h replacing^[[?25l^[[?25h multip^[[?25l^[[?25hles^[[
?25l^[[?25h of^[[?25l^[[?25h ^[[?25l^[[?25h3^[[?25l^[[?25h with^[[
?25l^[[?25h "^[[?25l^[[?25hF^[[?25l^[[?25hizz^[[?25l^[[?25h",^[[?2
5l^[[?25h multip^[[?25l^[[?25hles^[[?25l^[[?25h of^[[?25l^[[?25h ^
[[?25l^[[?25h5^[[?25l^[[?25h with^[[?25l^[[?25h
```
Over and over. If you try to do the `--json` output, it's no better. It would be nice to have the simple text output with nothing else.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6120/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6120/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8127
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8127/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8127/comments
|
https://api.github.com/repos/ollama/ollama/issues/8127/events
|
https://github.com/ollama/ollama/pull/8127
| 2,743,777,670
|
PR_kwDOJ0Z1Ps6FboqQ
| 8,127
|
llm: loosen format check to default to no format
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-12-17T02:21:37
| 2024-12-17T04:25:55
| 2024-12-17T02:45:47
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8127",
"html_url": "https://github.com/ollama/ollama/pull/8127",
"diff_url": "https://github.com/ollama/ollama/pull/8127.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8127.patch",
"merged_at": "2024-12-17T02:45:47"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8127/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8127/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1925
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1925/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1925/comments
|
https://api.github.com/repos/ollama/ollama/issues/1925/events
|
https://github.com/ollama/ollama/issues/1925
| 2,077,046,512
|
I_kwDOJ0Z1Ps57zTbw
| 1,925
|
Ollama is running in background in MacOS
|
{
"login": "MagzhanUnited",
"id": 123943870,
"node_id": "U_kgDOB2M7vg",
"avatar_url": "https://avatars.githubusercontent.com/u/123943870?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MagzhanUnited",
"html_url": "https://github.com/MagzhanUnited",
"followers_url": "https://api.github.com/users/MagzhanUnited/followers",
"following_url": "https://api.github.com/users/MagzhanUnited/following{/other_user}",
"gists_url": "https://api.github.com/users/MagzhanUnited/gists{/gist_id}",
"starred_url": "https://api.github.com/users/MagzhanUnited/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MagzhanUnited/subscriptions",
"organizations_url": "https://api.github.com/users/MagzhanUnited/orgs",
"repos_url": "https://api.github.com/users/MagzhanUnited/repos",
"events_url": "https://api.github.com/users/MagzhanUnited/events{/privacy}",
"received_events_url": "https://api.github.com/users/MagzhanUnited/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 11
| 2024-01-11T16:04:05
| 2025-01-29T08:51:22
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Even if I exit the ollama app I can see the ollama among run processes
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1925/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1925/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/1806
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1806/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1806/comments
|
https://api.github.com/repos/ollama/ollama/issues/1806/events
|
https://github.com/ollama/ollama/issues/1806
| 2,067,309,672
|
I_kwDOJ0Z1Ps57OKRo
| 1,806
|
Add cli switch to show generation time and tokens/sec output time
|
{
"login": "thawkins",
"id": 14537,
"node_id": "MDQ6VXNlcjE0NTM3",
"avatar_url": "https://avatars.githubusercontent.com/u/14537?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/thawkins",
"html_url": "https://github.com/thawkins",
"followers_url": "https://api.github.com/users/thawkins/followers",
"following_url": "https://api.github.com/users/thawkins/following{/other_user}",
"gists_url": "https://api.github.com/users/thawkins/gists{/gist_id}",
"starred_url": "https://api.github.com/users/thawkins/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/thawkins/subscriptions",
"organizations_url": "https://api.github.com/users/thawkins/orgs",
"repos_url": "https://api.github.com/users/thawkins/repos",
"events_url": "https://api.github.com/users/thawkins/events{/privacy}",
"received_events_url": "https://api.github.com/users/thawkins/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2024-01-05T12:56:05
| 2024-01-08T18:56:01
| 2024-01-08T18:56:01
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Would it be possible to add a metrics switch to show net generation time and output time with tokens/seconds. This would make comparing the performance of LLMs easier.
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1806/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1806/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4451
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4451/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4451/comments
|
https://api.github.com/repos/ollama/ollama/issues/4451/events
|
https://github.com/ollama/ollama/pull/4451
| 2,297,709,787
|
PR_kwDOJ0Z1Ps5vhZ4t
| 4,451
|
add ability to create a client without env file
|
{
"login": "Tyrell04",
"id": 43107913,
"node_id": "MDQ6VXNlcjQzMTA3OTEz",
"avatar_url": "https://avatars.githubusercontent.com/u/43107913?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tyrell04",
"html_url": "https://github.com/Tyrell04",
"followers_url": "https://api.github.com/users/Tyrell04/followers",
"following_url": "https://api.github.com/users/Tyrell04/following{/other_user}",
"gists_url": "https://api.github.com/users/Tyrell04/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Tyrell04/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Tyrell04/subscriptions",
"organizations_url": "https://api.github.com/users/Tyrell04/orgs",
"repos_url": "https://api.github.com/users/Tyrell04/repos",
"events_url": "https://api.github.com/users/Tyrell04/events{/privacy}",
"received_events_url": "https://api.github.com/users/Tyrell04/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2024-05-15T11:57:58
| 2024-05-19T04:40:18
| 2024-05-18T03:08:30
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4451",
"html_url": "https://github.com/ollama/ollama/pull/4451",
"diff_url": "https://github.com/ollama/ollama/pull/4451.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4451.patch",
"merged_at": null
}
|
If you use Ollama as a Go package, you are forced to use an environment variable as a configuration to change the host. With the CreateClient function, you are free to use your own configuration passed into the Ollama client.
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4451/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4451/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7324
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7324/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7324/comments
|
https://api.github.com/repos/ollama/ollama/issues/7324/events
|
https://github.com/ollama/ollama/pull/7324
| 2,606,172,201
|
PR_kwDOJ0Z1Ps5_f5eG
| 7,324
|
docs: use mingw32-make rather than MSYS Makefiles
|
{
"login": "shou692199",
"id": 45505768,
"node_id": "MDQ6VXNlcjQ1NTA1NzY4",
"avatar_url": "https://avatars.githubusercontent.com/u/45505768?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shou692199",
"html_url": "https://github.com/shou692199",
"followers_url": "https://api.github.com/users/shou692199/followers",
"following_url": "https://api.github.com/users/shou692199/following{/other_user}",
"gists_url": "https://api.github.com/users/shou692199/gists{/gist_id}",
"starred_url": "https://api.github.com/users/shou692199/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shou692199/subscriptions",
"organizations_url": "https://api.github.com/users/shou692199/orgs",
"repos_url": "https://api.github.com/users/shou692199/repos",
"events_url": "https://api.github.com/users/shou692199/events{/privacy}",
"received_events_url": "https://api.github.com/users/shou692199/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-10-22T18:26:14
| 2024-10-31T13:39:53
| 2024-10-31T13:39:53
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7324",
"html_url": "https://github.com/ollama/ollama/pull/7324",
"diff_url": "https://github.com/ollama/ollama/pull/7324.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7324.patch",
"merged_at": null
}
|
Since I'm trying to port ggml-sycl to llama package, MSYS Makefiles doesn't support icx compiler but MinGW Makefiles does, and it also works well with cpu, rocm and cuda. I think it's an ideal choice.
|
{
"login": "shou692199",
"id": 45505768,
"node_id": "MDQ6VXNlcjQ1NTA1NzY4",
"avatar_url": "https://avatars.githubusercontent.com/u/45505768?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shou692199",
"html_url": "https://github.com/shou692199",
"followers_url": "https://api.github.com/users/shou692199/followers",
"following_url": "https://api.github.com/users/shou692199/following{/other_user}",
"gists_url": "https://api.github.com/users/shou692199/gists{/gist_id}",
"starred_url": "https://api.github.com/users/shou692199/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shou692199/subscriptions",
"organizations_url": "https://api.github.com/users/shou692199/orgs",
"repos_url": "https://api.github.com/users/shou692199/repos",
"events_url": "https://api.github.com/users/shou692199/events{/privacy}",
"received_events_url": "https://api.github.com/users/shou692199/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7324/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7324/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2292
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2292/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2292/comments
|
https://api.github.com/repos/ollama/ollama/issues/2292/events
|
https://github.com/ollama/ollama/issues/2292
| 2,110,861,549
|
I_kwDOJ0Z1Ps590TDt
| 2,292
|
Manifest file?
|
{
"login": "EquaTechnologies",
"id": 107852330,
"node_id": "U_kgDOBm2yKg",
"avatar_url": "https://avatars.githubusercontent.com/u/107852330?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/EquaTechnologies",
"html_url": "https://github.com/EquaTechnologies",
"followers_url": "https://api.github.com/users/EquaTechnologies/followers",
"following_url": "https://api.github.com/users/EquaTechnologies/following{/other_user}",
"gists_url": "https://api.github.com/users/EquaTechnologies/gists{/gist_id}",
"starred_url": "https://api.github.com/users/EquaTechnologies/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/EquaTechnologies/subscriptions",
"organizations_url": "https://api.github.com/users/EquaTechnologies/orgs",
"repos_url": "https://api.github.com/users/EquaTechnologies/repos",
"events_url": "https://api.github.com/users/EquaTechnologies/events{/privacy}",
"received_events_url": "https://api.github.com/users/EquaTechnologies/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-01-31T19:57:51
| 2024-02-20T04:05:47
| 2024-02-20T04:05:47
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
qua@equa-Swift-SF314-54:~$ ollama run orca
pulling manifest
Error: pull model manifest: file does not exist
equa@equa-Swift-SF314-54:~$
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2292/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2292/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6240
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6240/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6240/comments
|
https://api.github.com/repos/ollama/ollama/issues/6240/events
|
https://github.com/ollama/ollama/issues/6240
| 2,454,100,567
|
I_kwDOJ0Z1Ps6SRppX
| 6,240
|
Not executed in gpu amd rx 6750 GRE
|
{
"login": "21307369",
"id": 47931342,
"node_id": "MDQ6VXNlcjQ3OTMxMzQy",
"avatar_url": "https://avatars.githubusercontent.com/u/47931342?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/21307369",
"html_url": "https://github.com/21307369",
"followers_url": "https://api.github.com/users/21307369/followers",
"following_url": "https://api.github.com/users/21307369/following{/other_user}",
"gists_url": "https://api.github.com/users/21307369/gists{/gist_id}",
"starred_url": "https://api.github.com/users/21307369/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/21307369/subscriptions",
"organizations_url": "https://api.github.com/users/21307369/orgs",
"repos_url": "https://api.github.com/users/21307369/repos",
"events_url": "https://api.github.com/users/21307369/events{/privacy}",
"received_events_url": "https://api.github.com/users/21307369/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] |
closed
| false
| null |
[] | null | 3
| 2024-08-07T18:39:39
| 2024-09-03T19:10:36
| 2024-08-09T19:25:52
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
<img width="898" alt="image" src="https://github.com/user-attachments/assets/87a706b4-7749-443f-86ea-86a3c7de1cc1">
2024/08/08 14:17:22 routes.go:1108: INFO server config env="map[OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_LLM_LIBRARY: OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/Users/lsmir2/.ollama/models OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[* http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://*] OLLAMA_RUNNERS_DIR: OLLAMA_SCHED_SPREAD:false OLLAMA_TMPDIR:]"
time=2024-08-08T14:17:22.106+08:00 level=INFO source=images.go:781 msg="total blobs: 5"
time=2024-08-08T14:17:22.106+08:00 level=INFO source=images.go:788 msg="total unused blobs removed: 0"
time=2024-08-08T14:17:22.107+08:00 level=INFO source=routes.go:1155 msg="Listening on 127.0.0.1:11434 (version 0.3.4)"
time=2024-08-08T14:17:22.113+08:00 level=WARN source=assets.go:100 msg="unable to cleanup stale tmpdir" path=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama1166271814 error="remove /var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama1166271814: directory not empty"
time=2024-08-08T14:17:22.114+08:00 level=WARN source=assets.go:100 msg="unable to cleanup stale tmpdir" path=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama2116948276 error="remove /var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama2116948276: directory not empty"
time=2024-08-08T14:17:22.114+08:00 level=WARN source=assets.go:100 msg="unable to cleanup stale tmpdir" path=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama3670480486 error="remove /var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama3670480486: directory not empty"
time=2024-08-08T14:17:22.115+08:00 level=INFO source=payload.go:30 msg="extracting embedded files" dir=/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama4115735691/runners
time=2024-08-08T14:17:22.146+08:00 level=INFO source=payload.go:44 msg="Dynamic LLM libraries [cpu cpu_avx cpu_avx2]"
time=2024-08-08T14:17:22.146+08:00 level=INFO source=types.go:105 msg="inference compute" id="" library=cpu compute="" driver=0.0 name="" total="32.0 GiB" available="16.1 GiB"
time=2024-08-08T14:17:24.808+08:00 level=INFO source=memory.go:309 msg="offload to cpu" layers.requested=-1 layers.model=41 layers.offload=0 layers.split="" memory.available="[16.1 GiB]" memory.required.full="5.7 GiB" memory.required.partial="0 B" memory.required.kv="320.0 MiB" memory.required.allocations="[5.7 GiB]" memory.weights.total="4.6 GiB" memory.weights.repeating="4.1 GiB" memory.weights.nonrepeating="485.6 MiB" memory.graph.full="561.0 MiB" memory.graph.partial="789.6 MiB"
time=2024-08-08T14:17:24.809+08:00 level=INFO source=server.go:392 msg="starting llama server" cmd="/var/folders/3j/0tc5g9350n128f02l1tm845m0000gn/T/ollama4115735691/runners/cpu_avx2/ollama_llama_server --model /Users/lsmir2/.ollama/models/blobs/sha256-816441b33390807d429fbdb1de7e33bb4d569ac68e2203bdbca5d8d79b5c7266 --ctx-size 8192 --batch-size 512 --embedding --log-disable --no-mmap --parallel 4 --port 57883"
time=2024-08-08T14:17:24.816+08:00 level=INFO source=sched.go:445 msg="loaded runners" count=1
time=2024-08-08T14:17:24.816+08:00 level=INFO source=server.go:592 msg="waiting for llama runner to start responding"
time=2024-08-08T14:17:24.817+08:00 level=INFO source=server.go:626 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=3535 commit="1e6f6554" tid="0x107f92600" timestamp=1723097844
INFO [main] system info | n_threads=6 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="0x107f92600" timestamp=1723097844 total_threads=6
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="6" port="57883" tid="0x107f92600" timestamp=1723097844
llama_model_loader: loaded meta data with 23 key-value pairs and 283 tensors from /Users/lsmir2/.ollama/models/blobs/sha256-816441b33390807d429fbdb1de7e33bb4d569ac68e2203bdbca5d8d79b5c7266 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = chatglm
llama_model_loader: - kv 1: general.name str = codegeex4-all-9b
llama_model_loader: - kv 2: chatglm.context_length u32 = 131072
llama_model_loader: - kv 3: chatglm.embedding_length u32 = 4096
llama_model_loader: - kv 4: chatglm.feed_forward_length u32 = 13696
llama_model_loader: - kv 5: chatglm.block_count u32 = 40
llama_model_loader: - kv 6: chatglm.attention.head_count u32 = 32
llama_model_loader: - kv 7: chatglm.attention.head_count_kv u32 = 2
llama_model_loader: - kv 8: chatglm.attention.layer_norm_rms_epsilon f32 = 0.000010
llama_model_loader: - kv 9: general.file_type u32 = 2
llama_model_loader: - kv 10: chatglm.rope.dimension_count u32 = 64
llama_model_loader: - kv 11: tokenizer.ggml.add_bos_token bool = false
llama_model_loader: - kv 12: chatglm.rope.freq_base f32 = 5000000.000000
llama_model_loader: - kv 13: tokenizer.ggml.model str = gpt2
llama_model_loader: - kv 14: tokenizer.ggml.pre str = chatglm-bpe
llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,151552] = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv 16: tokenizer.ggml.token_type arr[i32,151552] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv 17: tokenizer.ggml.merges arr[str,151073] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",...
llama_model_loader: - kv 18: tokenizer.ggml.padding_token_id u32 = 151329
llama_model_loader: - kv 19: tokenizer.ggml.eos_token_id u32 = 151329
llama_model_loader: - kv 20: tokenizer.ggml.eot_token_id u32 = 151336
llama_model_loader: - kv 21: tokenizer.ggml.unknown_token_id u32 = 151329
llama_model_loader: - kv 22: general.quantization_version u32 = 2
llama_model_loader: - type f32: 121 tensors
llama_model_loader: - type q4_0: 161 tensors
llama_model_loader: - type q6_K: 1 tensors
time=2024-08-08T14:17:25.069+08:00 level=INFO source=server.go:626 msg="waiting for server to become available" status="llm server loading model"
llm_load_vocab: special tokens cache size = 223
llm_load_vocab: token to piece cache size = 0.9732 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = chatglm
llm_load_print_meta: vocab type = BPE
llm_load_print_meta: n_vocab = 151552
llm_load_print_meta: n_merges = 151073
llm_load_print_meta: vocab_only = 0
llm_load_print_meta: n_ctx_train = 131072
llm_load_print_meta: n_embd = 4096
llm_load_print_meta: n_layer = 40
llm_load_print_meta: n_head = 32
llm_load_print_meta: n_head_kv = 2
llm_load_print_meta: n_rot = 64
llm_load_print_meta: n_swa = 0
llm_load_print_meta: n_embd_head_k = 128
llm_load_print_meta: n_embd_head_v = 128
llm_load_print_meta: n_gqa = 16
llm_load_print_meta: n_embd_k_gqa = 256
llm_load_print_meta: n_embd_v_gqa = 256
llm_load_print_meta: f_norm_eps = 0.0e+00
llm_load_print_meta: f_norm_rms_eps = 1.0e-05
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 13696
llm_load_print_meta: n_expert = 0
llm_load_print_meta: n_expert_used = 0
llm_load_print_meta: causal attn = 1
llm_load_print_meta: pooling type = 0
llm_load_print_meta: rope type = 0
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 5000000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_ctx_orig_yarn = 131072
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: model type = 9B
llm_load_print_meta: model ftype = Q4_0
llm_load_print_meta: model params = 9.40 B
llm_load_print_meta: model size = 5.08 GiB (4.64 BPW)
llm_load_print_meta: general.name = codegeex4-all-9b
llm_load_print_meta: EOS token = 151329 '<|endoftext|>'
llm_load_print_meta: UNK token = 151329 '<|endoftext|>'
llm_load_print_meta: PAD token = 151329 '<|endoftext|>'
llm_load_print_meta: LF token = 128 'Ä'
llm_load_print_meta: EOT token = 151336 '<|user|>'
llm_load_print_meta: max token length = 1024
llm_load_tensors: ggml ctx size = 0.14 MiB
llm_load_tensors: CPU buffer size = 5196.84 MiB
llama_new_context_with_model: n_ctx = 8192
llama_new_context_with_model: n_batch = 512
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 5000000.0
llama_new_context_with_model: freq_scale = 1
llama_kv_cache_init: CPU KV buffer size = 320.00 MiB
llama_new_context_with_model: KV self size = 320.00 MiB, K (f16): 160.00 MiB, V (f16): 160.00 MiB
llama_new_context_with_model: CPU output buffer size = 2.38 MiB
llama_new_context_with_model: CPU compute buffer size = 561.01 MiB
llama_new_context_with_model: graph nodes = 1606
llama_new_context_with_model: graph splits = 1
INFO [main] model loaded | tid="0x107f92600" timestamp=1723097848
time=2024-08-08T14:17:28.352+08:00 level=INFO source=server.go:631 msg="llama runner started in 3.54 seconds"
[GIN] 2024/08/08 - 14:17:33 | 200 | 8.515598533s | 127.0.0.1 | POST "/v1/chat/completions"
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6240/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6240/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5511
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5511/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5511/comments
|
https://api.github.com/repos/ollama/ollama/issues/5511/events
|
https://github.com/ollama/ollama/pull/5511
| 2,393,285,539
|
PR_kwDOJ0Z1Ps50kwRU
| 5,511
|
llm: fix missing dylibs by restoring old build behavior on Linux and macOS
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-07-06T01:00:34
| 2024-07-08T16:55:51
| 2024-07-06T01:48:32
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5511",
"html_url": "https://github.com/ollama/ollama/pull/5511",
"diff_url": "https://github.com/ollama/ollama/pull/5511.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5511.patch",
"merged_at": "2024-07-06T01:48:32"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5511/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5511/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7660
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7660/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7660/comments
|
https://api.github.com/repos/ollama/ollama/issues/7660/events
|
https://github.com/ollama/ollama/issues/7660
| 2,657,322,190
|
I_kwDOJ0Z1Ps6eY4TO
| 7,660
|
on ollama model specifics web sites , add comaptible copypaste <name>:<tag> text
|
{
"login": "fxmbsw7",
"id": 39368685,
"node_id": "MDQ6VXNlcjM5MzY4Njg1",
"avatar_url": "https://avatars.githubusercontent.com/u/39368685?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/fxmbsw7",
"html_url": "https://github.com/fxmbsw7",
"followers_url": "https://api.github.com/users/fxmbsw7/followers",
"following_url": "https://api.github.com/users/fxmbsw7/following{/other_user}",
"gists_url": "https://api.github.com/users/fxmbsw7/gists{/gist_id}",
"starred_url": "https://api.github.com/users/fxmbsw7/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/fxmbsw7/subscriptions",
"organizations_url": "https://api.github.com/users/fxmbsw7/orgs",
"repos_url": "https://api.github.com/users/fxmbsw7/repos",
"events_url": "https://api.github.com/users/fxmbsw7/events{/privacy}",
"received_events_url": "https://api.github.com/users/fxmbsw7/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] |
closed
| false
|
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 9
| 2024-11-14T02:16:13
| 2024-12-01T07:31:54
| 2024-11-15T02:24:09
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
eg , user visits page of a model
chooses select list , and selects some model version
there on that version of model page i want
say , for qwen2.5-coder.on its 14b-instruct-q4_K_M page
simply qwen2.5-coder:14b-instruct-q4_K_M
name:tag
like hf s use that button with ollama copypaste run cmd in
oh and yea , peeferably also a small ' copy me ' button left on it
greets
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7660/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7660/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/630
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/630/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/630/comments
|
https://api.github.com/repos/ollama/ollama/issues/630/events
|
https://github.com/ollama/ollama/issues/630
| 1,916,976,222
|
I_kwDOJ0Z1Ps5yQrxe
| 630
|
Error: failed to start a llama runner
|
{
"login": "azhang",
"id": 940275,
"node_id": "MDQ6VXNlcjk0MDI3NQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/940275?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/azhang",
"html_url": "https://github.com/azhang",
"followers_url": "https://api.github.com/users/azhang/followers",
"following_url": "https://api.github.com/users/azhang/following{/other_user}",
"gists_url": "https://api.github.com/users/azhang/gists{/gist_id}",
"starred_url": "https://api.github.com/users/azhang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/azhang/subscriptions",
"organizations_url": "https://api.github.com/users/azhang/orgs",
"repos_url": "https://api.github.com/users/azhang/repos",
"events_url": "https://api.github.com/users/azhang/events{/privacy}",
"received_events_url": "https://api.github.com/users/azhang/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 28
| 2023-09-28T08:25:32
| 2024-07-12T19:18:02
| 2023-10-12T15:16:38
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
When I run
ollama run mistral
it downloads properly but then fails to run it, with the following error:
Error: failed to start a llama runner
I'm running this on my intel mbp with 64g ram
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/630/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/630/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3661
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3661/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3661/comments
|
https://api.github.com/repos/ollama/ollama/issues/3661/events
|
https://github.com/ollama/ollama/pull/3661
| 2,244,821,343
|
PR_kwDOJ0Z1Ps5svQOm
| 3,661
|
better checking for OLLAMA_HOST variable
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-04-15T23:39:21
| 2024-04-29T23:14:08
| 2024-04-29T23:14:07
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3661",
"html_url": "https://github.com/ollama/ollama/pull/3661",
"diff_url": "https://github.com/ollama/ollama/pull/3661.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3661.patch",
"merged_at": "2024-04-29T23:14:07"
}
|
This change adds better validation to the `OLLAMA_HOST` variable when used with `ollama serve`. It should work with both IPv4 and IPv6, and includes unit tests.
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3661/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3661/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1109
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1109/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1109/comments
|
https://api.github.com/repos/ollama/ollama/issues/1109/events
|
https://github.com/ollama/ollama/issues/1109
| 1,990,754,546
|
I_kwDOJ0Z1Ps52qIDy
| 1,109
|
Exploring Multimodal LLMs: Incorporating Image as Input in ModelFile.
|
{
"login": "saifullah-N",
"id": 79261602,
"node_id": "MDQ6VXNlcjc5MjYxNjAy",
"avatar_url": "https://avatars.githubusercontent.com/u/79261602?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/saifullah-N",
"html_url": "https://github.com/saifullah-N",
"followers_url": "https://api.github.com/users/saifullah-N/followers",
"following_url": "https://api.github.com/users/saifullah-N/following{/other_user}",
"gists_url": "https://api.github.com/users/saifullah-N/gists{/gist_id}",
"starred_url": "https://api.github.com/users/saifullah-N/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/saifullah-N/subscriptions",
"organizations_url": "https://api.github.com/users/saifullah-N/orgs",
"repos_url": "https://api.github.com/users/saifullah-N/repos",
"events_url": "https://api.github.com/users/saifullah-N/events{/privacy}",
"received_events_url": "https://api.github.com/users/saifullah-N/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2023-11-13T14:18:07
| 2023-11-13T17:42:57
| 2023-11-13T17:42:57
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi, I'd like to experiment with multimodal language models that can handle both images and text as input. Is there a way to input an image.
My ModelFile looks like this
```
FROM ./my_model.gguf
PARAMERTER..
SYSTEM You are An Artist describe the image
```
I'm trying to run this on macos.
M1 chip
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1109/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1109/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3728
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3728/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3728/comments
|
https://api.github.com/repos/ollama/ollama/issues/3728/events
|
https://github.com/ollama/ollama/issues/3728
| 2,249,868,609
|
I_kwDOJ0Z1Ps6GGkVB
| 3,728
|
ollama prune
|
{
"login": "Tobiadefami",
"id": 48369656,
"node_id": "MDQ6VXNlcjQ4MzY5NjU2",
"avatar_url": "https://avatars.githubusercontent.com/u/48369656?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tobiadefami",
"html_url": "https://github.com/Tobiadefami",
"followers_url": "https://api.github.com/users/Tobiadefami/followers",
"following_url": "https://api.github.com/users/Tobiadefami/following{/other_user}",
"gists_url": "https://api.github.com/users/Tobiadefami/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Tobiadefami/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Tobiadefami/subscriptions",
"organizations_url": "https://api.github.com/users/Tobiadefami/orgs",
"repos_url": "https://api.github.com/users/Tobiadefami/repos",
"events_url": "https://api.github.com/users/Tobiadefami/events{/privacy}",
"received_events_url": "https://api.github.com/users/Tobiadefami/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-04-18T06:35:53
| 2024-04-18T08:19:21
| 2024-04-18T08:19:21
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I think ollama needs a prune command, something similar to `docker system prune`. Just to keep the environment tidy
|
{
"login": "Tobiadefami",
"id": 48369656,
"node_id": "MDQ6VXNlcjQ4MzY5NjU2",
"avatar_url": "https://avatars.githubusercontent.com/u/48369656?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Tobiadefami",
"html_url": "https://github.com/Tobiadefami",
"followers_url": "https://api.github.com/users/Tobiadefami/followers",
"following_url": "https://api.github.com/users/Tobiadefami/following{/other_user}",
"gists_url": "https://api.github.com/users/Tobiadefami/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Tobiadefami/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Tobiadefami/subscriptions",
"organizations_url": "https://api.github.com/users/Tobiadefami/orgs",
"repos_url": "https://api.github.com/users/Tobiadefami/repos",
"events_url": "https://api.github.com/users/Tobiadefami/events{/privacy}",
"received_events_url": "https://api.github.com/users/Tobiadefami/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3728/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3728/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3660
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3660/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3660/comments
|
https://api.github.com/repos/ollama/ollama/issues/3660/events
|
https://github.com/ollama/ollama/issues/3660
| 2,244,762,677
|
I_kwDOJ0Z1Ps6FzFw1
| 3,660
|
Exception using command-r-plus:104b-q4_0
|
{
"login": "mehdiataei",
"id": 29686241,
"node_id": "MDQ6VXNlcjI5Njg2MjQx",
"avatar_url": "https://avatars.githubusercontent.com/u/29686241?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mehdiataei",
"html_url": "https://github.com/mehdiataei",
"followers_url": "https://api.github.com/users/mehdiataei/followers",
"following_url": "https://api.github.com/users/mehdiataei/following{/other_user}",
"gists_url": "https://api.github.com/users/mehdiataei/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mehdiataei/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mehdiataei/subscriptions",
"organizations_url": "https://api.github.com/users/mehdiataei/orgs",
"repos_url": "https://api.github.com/users/mehdiataei/repos",
"events_url": "https://api.github.com/users/mehdiataei/events{/privacy}",
"received_events_url": "https://api.github.com/users/mehdiataei/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 7
| 2024-04-15T22:49:24
| 2024-04-17T00:41:10
| 2024-04-17T00:41:10
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Ollama: exception done_getting_tensors: wrong number of tensors; expected 642, got 514
### What did you expect to see?
_No response_
### Steps to reproduce
_No response_
### Are there any recent changes that introduced the issue?
_No response_
### OS
_No response_
### Architecture
_No response_
### Platform
_No response_
### Ollama version
_No response_
### GPU
_No response_
### GPU info
_No response_
### CPU
_No response_
### Other software
_No response_
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3660/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3660/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2792
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2792/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2792/comments
|
https://api.github.com/repos/ollama/ollama/issues/2792/events
|
https://github.com/ollama/ollama/issues/2792
| 2,157,751,613
|
I_kwDOJ0Z1Ps6AnK09
| 2,792
|
Subsequent generation requests hang after successful generation request with `num_predict: 0`
|
{
"login": "stanier",
"id": 8770519,
"node_id": "MDQ6VXNlcjg3NzA1MTk=",
"avatar_url": "https://avatars.githubusercontent.com/u/8770519?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/stanier",
"html_url": "https://github.com/stanier",
"followers_url": "https://api.github.com/users/stanier/followers",
"following_url": "https://api.github.com/users/stanier/following{/other_user}",
"gists_url": "https://api.github.com/users/stanier/gists{/gist_id}",
"starred_url": "https://api.github.com/users/stanier/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/stanier/subscriptions",
"organizations_url": "https://api.github.com/users/stanier/orgs",
"repos_url": "https://api.github.com/users/stanier/repos",
"events_url": "https://api.github.com/users/stanier/events{/privacy}",
"received_events_url": "https://api.github.com/users/stanier/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-02-27T22:54:24
| 2024-05-10T01:13:35
| 2024-05-10T01:13:35
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
If you submit a generation request with `num_predict: 0`, the request will be handled successfully but all subsequent generation requests will hang indefinitely regardless of their `num_predict` values.
Below is an example of how to reproduce this behavior:
```
curl http://localhost:11434/api/generate -d '{
"model": "vicuna:13b-16k",
"template":"Hello world!",
"stream":false,
"num_predict":0
}'
```
Any requests following this will not return until the daemon has been restarted.
Also to note is that the daemon might be in an unresponsive state following this request, I've had to `kill` it each time it's happened to me so far, but haven't tested this against typical behavior for signals sent mid-generation or otherwise.
I think it also affects `raw` mode, and neither `num_keep` or `num_ctx` seem to be related.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2792/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2792/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1849
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1849/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1849/comments
|
https://api.github.com/repos/ollama/ollama/issues/1849/events
|
https://github.com/ollama/ollama/pull/1849
| 2,069,374,142
|
PR_kwDOJ0Z1Ps5jatea
| 1,849
|
Accomodate split cuda lib dir
|
{
"login": "mraiser",
"id": 1268027,
"node_id": "MDQ6VXNlcjEyNjgwMjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/1268027?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mraiser",
"html_url": "https://github.com/mraiser",
"followers_url": "https://api.github.com/users/mraiser/followers",
"following_url": "https://api.github.com/users/mraiser/following{/other_user}",
"gists_url": "https://api.github.com/users/mraiser/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mraiser/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mraiser/subscriptions",
"organizations_url": "https://api.github.com/users/mraiser/orgs",
"repos_url": "https://api.github.com/users/mraiser/repos",
"events_url": "https://api.github.com/users/mraiser/events{/privacy}",
"received_events_url": "https://api.github.com/users/mraiser/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 9
| 2024-01-08T00:24:45
| 2024-02-06T00:01:17
| 2024-02-06T00:01:17
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1849",
"html_url": "https://github.com/ollama/ollama/pull/1849",
"diff_url": "https://github.com/ollama/ollama/pull/1849.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1849.patch",
"merged_at": "2024-02-06T00:01:17"
}
|
Makes it a little easier to compile when cuda lib dir is split up as in nixos.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1849/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1849/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6786
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6786/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6786/comments
|
https://api.github.com/repos/ollama/ollama/issues/6786/events
|
https://github.com/ollama/ollama/issues/6786
| 2,523,912,567
|
I_kwDOJ0Z1Ps6Wb9l3
| 6,786
|
Isn't it time to move onto Omni models?
|
{
"login": "Meshwa428",
"id": 135232056,
"node_id": "U_kgDOCA96OA",
"avatar_url": "https://avatars.githubusercontent.com/u/135232056?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Meshwa428",
"html_url": "https://github.com/Meshwa428",
"followers_url": "https://api.github.com/users/Meshwa428/followers",
"following_url": "https://api.github.com/users/Meshwa428/following{/other_user}",
"gists_url": "https://api.github.com/users/Meshwa428/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Meshwa428/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Meshwa428/subscriptions",
"organizations_url": "https://api.github.com/users/Meshwa428/orgs",
"repos_url": "https://api.github.com/users/Meshwa428/repos",
"events_url": "https://api.github.com/users/Meshwa428/events{/privacy}",
"received_events_url": "https://api.github.com/users/Meshwa428/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-09-13T05:34:46
| 2024-10-23T17:17:40
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
There is a model that I found today called
*LLaMa 3.1 8b Omni* and it is a speech to speech model with very low latency ensuring the best experience for local models.
But as for Ollama it doesn't support such models, although VLMs are there but having these Omni models on your local device is just 🤌🏻 too Good to be true.
HF reference for llama 3.1 8b Omni:
https://huggingface.co/ICTNLP/Llama-3.1-8B-Omni
Looking forward to @ollama team's implementation
Thanks🙏🏻
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6786/reactions",
"total_count": 9,
"+1": 9,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6786/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/5244
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5244/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5244/comments
|
https://api.github.com/repos/ollama/ollama/issues/5244/events
|
https://github.com/ollama/ollama/pull/5244
| 2,368,887,556
|
PR_kwDOJ0Z1Ps5zS5Hb
| 5,244
|
llm: suppress large allocations for GGUF arrays
|
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers",
"following_url": "https://api.github.com/users/bmizerany/following{/other_user}",
"gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions",
"organizations_url": "https://api.github.com/users/bmizerany/orgs",
"repos_url": "https://api.github.com/users/bmizerany/repos",
"events_url": "https://api.github.com/users/bmizerany/events{/privacy}",
"received_events_url": "https://api.github.com/users/bmizerany/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-06-23T21:01:06
| 2024-06-23T22:32:04
| 2024-06-23T22:32:04
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | true
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5244",
"html_url": "https://github.com/ollama/ollama/pull/5244",
"diff_url": "https://github.com/ollama/ollama/pull/5244.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5244.patch",
"merged_at": null
}
|
This introduces a little array type for holding GGUF arrays that prevents the array from growing too large. It preserves the total size of the array, but limits the number of elements that are actually allocated.
GGUF arrays that are extremely large, such as tokens, etc, are generally uninteresting to users, and are not worth the memory overhead, and the time spent allocating and freeing them. They are necessary for inference, but not for inspection.
The size of these arrays is, however, important in Ollama, so it is preserved in a separate field on array.
|
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers",
"following_url": "https://api.github.com/users/bmizerany/following{/other_user}",
"gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions",
"organizations_url": "https://api.github.com/users/bmizerany/orgs",
"repos_url": "https://api.github.com/users/bmizerany/repos",
"events_url": "https://api.github.com/users/bmizerany/events{/privacy}",
"received_events_url": "https://api.github.com/users/bmizerany/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5244/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5244/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/836
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/836/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/836/comments
|
https://api.github.com/repos/ollama/ollama/issues/836/events
|
https://github.com/ollama/ollama/issues/836
| 1,949,521,611
|
I_kwDOJ0Z1Ps50M1bL
| 836
|
Support setting parameters without a modelfile
|
{
"login": "chigkim",
"id": 22120994,
"node_id": "MDQ6VXNlcjIyMTIwOTk0",
"avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chigkim",
"html_url": "https://github.com/chigkim",
"followers_url": "https://api.github.com/users/chigkim/followers",
"following_url": "https://api.github.com/users/chigkim/following{/other_user}",
"gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}",
"starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/chigkim/subscriptions",
"organizations_url": "https://api.github.com/users/chigkim/orgs",
"repos_url": "https://api.github.com/users/chigkim/repos",
"events_url": "https://api.github.com/users/chigkim/events{/privacy}",
"received_events_url": "https://api.github.com/users/chigkim/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3
| 2023-10-18T11:30:51
| 2024-05-05T23:14:26
| 2024-01-03T03:38:43
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I know you can set temperature using modelfile or api, but it would be much more convenient to be able to do it in cli.
For example:
`ollama run llama-2 -temperture 0.1` when loading
and/or
`/set temperature 0.1` in runtime
Thanks for your consideration!
|
{
"login": "chigkim",
"id": 22120994,
"node_id": "MDQ6VXNlcjIyMTIwOTk0",
"avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chigkim",
"html_url": "https://github.com/chigkim",
"followers_url": "https://api.github.com/users/chigkim/followers",
"following_url": "https://api.github.com/users/chigkim/following{/other_user}",
"gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}",
"starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/chigkim/subscriptions",
"organizations_url": "https://api.github.com/users/chigkim/orgs",
"repos_url": "https://api.github.com/users/chigkim/repos",
"events_url": "https://api.github.com/users/chigkim/events{/privacy}",
"received_events_url": "https://api.github.com/users/chigkim/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/836/reactions",
"total_count": 5,
"+1": 5,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/836/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4747
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4747/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4747/comments
|
https://api.github.com/repos/ollama/ollama/issues/4747/events
|
https://github.com/ollama/ollama/issues/4747
| 2,327,194,470
|
I_kwDOJ0Z1Ps6Ktitm
| 4,747
|
Running multiple models simultaneously, always using one card
|
{
"login": "leoHostProject",
"id": 87935281,
"node_id": "MDQ6VXNlcjg3OTM1Mjgx",
"avatar_url": "https://avatars.githubusercontent.com/u/87935281?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/leoHostProject",
"html_url": "https://github.com/leoHostProject",
"followers_url": "https://api.github.com/users/leoHostProject/followers",
"following_url": "https://api.github.com/users/leoHostProject/following{/other_user}",
"gists_url": "https://api.github.com/users/leoHostProject/gists{/gist_id}",
"starred_url": "https://api.github.com/users/leoHostProject/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/leoHostProject/subscriptions",
"organizations_url": "https://api.github.com/users/leoHostProject/orgs",
"repos_url": "https://api.github.com/users/leoHostProject/repos",
"events_url": "https://api.github.com/users/leoHostProject/events{/privacy}",
"received_events_url": "https://api.github.com/users/leoHostProject/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-05-31T07:58:11
| 2024-05-31T18:58:02
| 2024-05-31T18:57:10
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Running multiple models simultaneously, always using one card,but i have 4 cards and download 4model
When multiple users are using it at the same time, always clear the first card and then load other models instead of using my other idle cards
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.33
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4747/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4747/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4583
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4583/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4583/comments
|
https://api.github.com/repos/ollama/ollama/issues/4583/events
|
https://github.com/ollama/ollama/pull/4583
| 2,311,914,296
|
PR_kwDOJ0Z1Ps5wRyOi
| 4,583
|
Add new community integration (TypingMind)
|
{
"login": "trungdq88",
"id": 4214509,
"node_id": "MDQ6VXNlcjQyMTQ1MDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4214509?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trungdq88",
"html_url": "https://github.com/trungdq88",
"followers_url": "https://api.github.com/users/trungdq88/followers",
"following_url": "https://api.github.com/users/trungdq88/following{/other_user}",
"gists_url": "https://api.github.com/users/trungdq88/gists{/gist_id}",
"starred_url": "https://api.github.com/users/trungdq88/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/trungdq88/subscriptions",
"organizations_url": "https://api.github.com/users/trungdq88/orgs",
"repos_url": "https://api.github.com/users/trungdq88/repos",
"events_url": "https://api.github.com/users/trungdq88/events{/privacy}",
"received_events_url": "https://api.github.com/users/trungdq88/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-05-23T04:31:39
| 2024-06-10T09:27:58
| 2024-06-10T09:27:57
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4583",
"html_url": "https://github.com/ollama/ollama/pull/4583",
"diff_url": "https://github.com/ollama/ollama/pull/4583.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4583.patch",
"merged_at": null
}
|
Cheers!
|
{
"login": "trungdq88",
"id": 4214509,
"node_id": "MDQ6VXNlcjQyMTQ1MDk=",
"avatar_url": "https://avatars.githubusercontent.com/u/4214509?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/trungdq88",
"html_url": "https://github.com/trungdq88",
"followers_url": "https://api.github.com/users/trungdq88/followers",
"following_url": "https://api.github.com/users/trungdq88/following{/other_user}",
"gists_url": "https://api.github.com/users/trungdq88/gists{/gist_id}",
"starred_url": "https://api.github.com/users/trungdq88/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/trungdq88/subscriptions",
"organizations_url": "https://api.github.com/users/trungdq88/orgs",
"repos_url": "https://api.github.com/users/trungdq88/repos",
"events_url": "https://api.github.com/users/trungdq88/events{/privacy}",
"received_events_url": "https://api.github.com/users/trungdq88/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4583/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4583/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5688
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5688/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5688/comments
|
https://api.github.com/repos/ollama/ollama/issues/5688/events
|
https://github.com/ollama/ollama/pull/5688
| 2,407,318,220
|
PR_kwDOJ0Z1Ps51T_VY
| 5,688
|
Serve static files
|
{
"login": "1feralcat",
"id": 51179976,
"node_id": "MDQ6VXNlcjUxMTc5OTc2",
"avatar_url": "https://avatars.githubusercontent.com/u/51179976?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/1feralcat",
"html_url": "https://github.com/1feralcat",
"followers_url": "https://api.github.com/users/1feralcat/followers",
"following_url": "https://api.github.com/users/1feralcat/following{/other_user}",
"gists_url": "https://api.github.com/users/1feralcat/gists{/gist_id}",
"starred_url": "https://api.github.com/users/1feralcat/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/1feralcat/subscriptions",
"organizations_url": "https://api.github.com/users/1feralcat/orgs",
"repos_url": "https://api.github.com/users/1feralcat/repos",
"events_url": "https://api.github.com/users/1feralcat/events{/privacy}",
"received_events_url": "https://api.github.com/users/1feralcat/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-07-14T08:06:08
| 2024-08-14T16:46:19
| 2024-08-14T16:46:19
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5688",
"html_url": "https://github.com/ollama/ollama/pull/5688",
"diff_url": "https://github.com/ollama/ollama/pull/5688.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5688.patch",
"merged_at": null
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5688/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5688/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5707
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5707/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5707/comments
|
https://api.github.com/repos/ollama/ollama/issues/5707/events
|
https://github.com/ollama/ollama/pull/5707
| 2,409,438,713
|
PR_kwDOJ0Z1Ps51bLDN
| 5,707
|
Image Embeddings
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-07-15T19:13:38
| 2024-11-21T10:11:28
| 2024-11-21T10:11:28
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | true
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5707",
"html_url": "https://github.com/ollama/ollama/pull/5707",
"diff_url": "https://github.com/ollama/ollama/pull/5707.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5707.patch",
"merged_at": null
}
|
rough work
|
{
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/users/mchiang0610/followers",
"following_url": "https://api.github.com/users/mchiang0610/following{/other_user}",
"gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions",
"organizations_url": "https://api.github.com/users/mchiang0610/orgs",
"repos_url": "https://api.github.com/users/mchiang0610/repos",
"events_url": "https://api.github.com/users/mchiang0610/events{/privacy}",
"received_events_url": "https://api.github.com/users/mchiang0610/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5707/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5707/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6117
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6117/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6117/comments
|
https://api.github.com/repos/ollama/ollama/issues/6117/events
|
https://github.com/ollama/ollama/issues/6117
| 2,442,322,095
|
I_kwDOJ0Z1Ps6RkuCv
| 6,117
|
Add Gemma 2 2b base/ text/ pre-trained model to registry
|
{
"login": "nviraj",
"id": 8409854,
"node_id": "MDQ6VXNlcjg0MDk4NTQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/8409854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nviraj",
"html_url": "https://github.com/nviraj",
"followers_url": "https://api.github.com/users/nviraj/followers",
"following_url": "https://api.github.com/users/nviraj/following{/other_user}",
"gists_url": "https://api.github.com/users/nviraj/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nviraj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nviraj/subscriptions",
"organizations_url": "https://api.github.com/users/nviraj/orgs",
"repos_url": "https://api.github.com/users/nviraj/repos",
"events_url": "https://api.github.com/users/nviraj/events{/privacy}",
"received_events_url": "https://api.github.com/users/nviraj/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-08-01T12:25:31
| 2024-08-02T12:19:55
| 2024-08-02T12:19:55
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Apologies, if I am missing something but can you upload quants for [base 2b](https://huggingface.co/google/gemma-2-2b) as well?
All I see are instruct ones [here](https://ollama.com/library/gemma2/tags)
Thanks!
|
{
"login": "nviraj",
"id": 8409854,
"node_id": "MDQ6VXNlcjg0MDk4NTQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/8409854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nviraj",
"html_url": "https://github.com/nviraj",
"followers_url": "https://api.github.com/users/nviraj/followers",
"following_url": "https://api.github.com/users/nviraj/following{/other_user}",
"gists_url": "https://api.github.com/users/nviraj/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nviraj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nviraj/subscriptions",
"organizations_url": "https://api.github.com/users/nviraj/orgs",
"repos_url": "https://api.github.com/users/nviraj/repos",
"events_url": "https://api.github.com/users/nviraj/events{/privacy}",
"received_events_url": "https://api.github.com/users/nviraj/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6117/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6117/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4652
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4652/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4652/comments
|
https://api.github.com/repos/ollama/ollama/issues/4652/events
|
https://github.com/ollama/ollama/pull/4652
| 2,317,888,554
|
PR_kwDOJ0Z1Ps5wmNS4
| 4,652
|
Ensure `nvidia` and `nvidia_uvm` kernel modules are loaded in `install.sh` script and at startup
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-05-26T18:53:36
| 2024-05-26T21:57:18
| 2024-05-26T21:57:17
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4652",
"html_url": "https://github.com/ollama/ollama/pull/4652",
"diff_url": "https://github.com/ollama/ollama/pull/4652.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4652.patch",
"merged_at": "2024-05-26T21:57:17"
}
|
Make sure the `nvidia` and `nvidia_uvm` kernel modules are loaded on install.
Nvidia has a daemon that takes care of that `nvidia-persistenced` for restarts, so add both to its config file as well
Lastly, make sure the kernel modules are loaded when re-running the install script, even if drivers are already installed.
Fixes https://github.com/ollama/ollama/issues/4563
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4652/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4652/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1988
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1988/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1988/comments
|
https://api.github.com/repos/ollama/ollama/issues/1988/events
|
https://github.com/ollama/ollama/pull/1988
| 2,080,801,827
|
PR_kwDOJ0Z1Ps5kBthD
| 1,988
|
Fix typo in arm mac arch script
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-01-14T16:34:14
| 2024-01-14T16:45:25
| 2024-01-14T16:45:18
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1988",
"html_url": "https://github.com/ollama/ollama/pull/1988",
"diff_url": "https://github.com/ollama/ollama/pull/1988.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1988.patch",
"merged_at": "2024-01-14T16:45:18"
}
| null |
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1988/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1988/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3401
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3401/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3401/comments
|
https://api.github.com/repos/ollama/ollama/issues/3401/events
|
https://github.com/ollama/ollama/issues/3401
| 2,214,382,654
|
I_kwDOJ0Z1Ps6D_Mw-
| 3,401
|
Document `OLLAMA_DEBUG` in `ollama serve` `-h` docs
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 5667396210,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2acg",
"url": "https://api.github.com/repos/ollama/ollama/labels/good%20first%20issue",
"name": "good first issue",
"color": "7057ff",
"default": true,
"description": "Good for newcomers"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-03-29T00:27:47
| 2024-05-13T20:20:58
| 2024-05-13T20:20:58
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What are you trying to do?
_No response_
### How should we solve this?
_No response_
### What is the impact of not solving this?
_No response_
### Anything else?
_No response_
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3401/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3401/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1648
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1648/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1648/comments
|
https://api.github.com/repos/ollama/ollama/issues/1648/events
|
https://github.com/ollama/ollama/issues/1648
| 2,051,491,566
|
I_kwDOJ0Z1Ps56R0bu
| 1,648
|
gpu issues
|
{
"login": "RootnuII",
"id": 66104474,
"node_id": "MDQ6VXNlcjY2MTA0NDc0",
"avatar_url": "https://avatars.githubusercontent.com/u/66104474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RootnuII",
"html_url": "https://github.com/RootnuII",
"followers_url": "https://api.github.com/users/RootnuII/followers",
"following_url": "https://api.github.com/users/RootnuII/following{/other_user}",
"gists_url": "https://api.github.com/users/RootnuII/gists{/gist_id}",
"starred_url": "https://api.github.com/users/RootnuII/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/RootnuII/subscriptions",
"organizations_url": "https://api.github.com/users/RootnuII/orgs",
"repos_url": "https://api.github.com/users/RootnuII/repos",
"events_url": "https://api.github.com/users/RootnuII/events{/privacy}",
"received_events_url": "https://api.github.com/users/RootnuII/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 5
| 2023-12-21T00:38:19
| 2023-12-23T15:16:26
| 2023-12-23T09:59:42
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
warning: gpu support may not be enabled, check that you have installed GPU drivers: nvidia-smi command failed
i have installed all drivers for my gpu:1070ti
still nothing maybe i installed the wrong drivers can someone point me in the right direction
|
{
"login": "RootnuII",
"id": 66104474,
"node_id": "MDQ6VXNlcjY2MTA0NDc0",
"avatar_url": "https://avatars.githubusercontent.com/u/66104474?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RootnuII",
"html_url": "https://github.com/RootnuII",
"followers_url": "https://api.github.com/users/RootnuII/followers",
"following_url": "https://api.github.com/users/RootnuII/following{/other_user}",
"gists_url": "https://api.github.com/users/RootnuII/gists{/gist_id}",
"starred_url": "https://api.github.com/users/RootnuII/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/RootnuII/subscriptions",
"organizations_url": "https://api.github.com/users/RootnuII/orgs",
"repos_url": "https://api.github.com/users/RootnuII/repos",
"events_url": "https://api.github.com/users/RootnuII/events{/privacy}",
"received_events_url": "https://api.github.com/users/RootnuII/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1648/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1648/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2828
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2828/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2828/comments
|
https://api.github.com/repos/ollama/ollama/issues/2828/events
|
https://github.com/ollama/ollama/pull/2828
| 2,160,642,796
|
PR_kwDOJ0Z1Ps5oQuDE
| 2,828
|
update llama.cpp submodule commit to `87c91c07`
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-29T07:59:29
| 2024-02-29T17:42:09
| 2024-02-29T17:42:08
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2828",
"html_url": "https://github.com/ollama/ollama/pull/2828",
"diff_url": "https://github.com/ollama/ollama/pull/2828.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2828.patch",
"merged_at": "2024-02-29T17:42:08"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2828/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2828/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1501
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1501/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1501/comments
|
https://api.github.com/repos/ollama/ollama/issues/1501/events
|
https://github.com/ollama/ollama/issues/1501
| 2,039,679,106
|
I_kwDOJ0Z1Ps55kwiC
| 1,501
|
macOS environment variable not working
|
{
"login": "brandoncarl",
"id": 5427971,
"node_id": "MDQ6VXNlcjU0Mjc5NzE=",
"avatar_url": "https://avatars.githubusercontent.com/u/5427971?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/brandoncarl",
"html_url": "https://github.com/brandoncarl",
"followers_url": "https://api.github.com/users/brandoncarl/followers",
"following_url": "https://api.github.com/users/brandoncarl/following{/other_user}",
"gists_url": "https://api.github.com/users/brandoncarl/gists{/gist_id}",
"starred_url": "https://api.github.com/users/brandoncarl/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/brandoncarl/subscriptions",
"organizations_url": "https://api.github.com/users/brandoncarl/orgs",
"repos_url": "https://api.github.com/users/brandoncarl/repos",
"events_url": "https://api.github.com/users/brandoncarl/events{/privacy}",
"received_events_url": "https://api.github.com/users/brandoncarl/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 9
| 2023-12-13T13:08:29
| 2025-01-11T11:41:21
| 2024-01-22T23:41:41
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello - thanks for the great repository. I wanted to alert you to the fact that the OLLAMA_MODELS path appears to be having no impact. This is true for `pull`, `rull` and serving.
```
$ echo $OLLAMA_MODELS
(prints appropriate directory)
$ ollama run <model>
(downloads to ~/.ollama/..)
$ OLLAMA_MODELS=<directory> run <model>
(downloads to ~/.ollama/...)
```
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1501/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1501/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4947
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4947/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4947/comments
|
https://api.github.com/repos/ollama/ollama/issues/4947/events
|
https://github.com/ollama/ollama/issues/4947
| 2,342,185,507
|
I_kwDOJ0Z1Ps6Lmuoj
| 4,947
|
Cannot update ollama on windows through "restart to update"
|
{
"login": "TonyBlur",
"id": 101348912,
"node_id": "U_kgDOBgp2MA",
"avatar_url": "https://avatars.githubusercontent.com/u/101348912?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TonyBlur",
"html_url": "https://github.com/TonyBlur",
"followers_url": "https://api.github.com/users/TonyBlur/followers",
"following_url": "https://api.github.com/users/TonyBlur/following{/other_user}",
"gists_url": "https://api.github.com/users/TonyBlur/gists{/gist_id}",
"starred_url": "https://api.github.com/users/TonyBlur/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/TonyBlur/subscriptions",
"organizations_url": "https://api.github.com/users/TonyBlur/orgs",
"repos_url": "https://api.github.com/users/TonyBlur/repos",
"events_url": "https://api.github.com/users/TonyBlur/events{/privacy}",
"received_events_url": "https://api.github.com/users/TonyBlur/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 11
| 2024-06-09T11:52:04
| 2024-06-10T11:42:01
| 2024-06-10T11:42:01
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
cannot update ollama on windows through clicking "restart to update" at taskbar, it has no response
### OS
Windows
### GPU
AMD
### CPU
AMD
### Ollama version
0.1.41
|
{
"login": "TonyBlur",
"id": 101348912,
"node_id": "U_kgDOBgp2MA",
"avatar_url": "https://avatars.githubusercontent.com/u/101348912?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TonyBlur",
"html_url": "https://github.com/TonyBlur",
"followers_url": "https://api.github.com/users/TonyBlur/followers",
"following_url": "https://api.github.com/users/TonyBlur/following{/other_user}",
"gists_url": "https://api.github.com/users/TonyBlur/gists{/gist_id}",
"starred_url": "https://api.github.com/users/TonyBlur/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/TonyBlur/subscriptions",
"organizations_url": "https://api.github.com/users/TonyBlur/orgs",
"repos_url": "https://api.github.com/users/TonyBlur/repos",
"events_url": "https://api.github.com/users/TonyBlur/events{/privacy}",
"received_events_url": "https://api.github.com/users/TonyBlur/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4947/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4947/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5918
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5918/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5918/comments
|
https://api.github.com/repos/ollama/ollama/issues/5918/events
|
https://github.com/ollama/ollama/issues/5918
| 2,427,947,507
|
I_kwDOJ0Z1Ps6Qt4nz
| 5,918
|
Llama3.1 70b-instruct-q4_1 buggy
|
{
"login": "velaia",
"id": 1515904,
"node_id": "MDQ6VXNlcjE1MTU5MDQ=",
"avatar_url": "https://avatars.githubusercontent.com/u/1515904?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/velaia",
"html_url": "https://github.com/velaia",
"followers_url": "https://api.github.com/users/velaia/followers",
"following_url": "https://api.github.com/users/velaia/following{/other_user}",
"gists_url": "https://api.github.com/users/velaia/gists{/gist_id}",
"starred_url": "https://api.github.com/users/velaia/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/velaia/subscriptions",
"organizations_url": "https://api.github.com/users/velaia/orgs",
"repos_url": "https://api.github.com/users/velaia/repos",
"events_url": "https://api.github.com/users/velaia/events{/privacy}",
"received_events_url": "https://api.github.com/users/velaia/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 14
| 2024-07-24T15:54:38
| 2024-09-15T17:35:22
| 2024-09-15T17:18:31
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
When I run the **70b-instruct-q4_1** version of Llama3.1 ollama gives a buggy reply:
My sample request:
> ➜ ollama-tests curl http://localhost:11434/api/chat -d '{
"model": "llama3.1:70b-instruct-q4_1",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Tell me about the top 3 commercial industrial computer vision products on the market."
}
], "stream": false
}'
The model's response
> {
"model": "llama3.1:70b-instruct-q4_1",
"created_at": "2024-07-24T15:45:23.026538Z",
"message": {
"role": "assistant",
"content": "assistant\nassistantassistantassistant"
},
"done_reason": "stop",
"done": true,
"total_duration": 56814340833,
"load_duration": 53209095791,
"prompt_eval_count": 37,
"prompt_eval_duration": 2107466000,
"eval_count": 6,
"eval_duration": 1493717000
}`
The same request generates a proper response using **llama3.1:8b-instruct-q8_0**.
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.2.8
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5918/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5918/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/2496
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2496/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2496/comments
|
https://api.github.com/repos/ollama/ollama/issues/2496/events
|
https://github.com/ollama/ollama/issues/2496
| 2,134,803,651
|
I_kwDOJ0Z1Ps5_PoTD
| 2,496
|
default num_thread incorrect on some large core count system (non-hyperthreading)
|
{
"login": "mokkin",
"id": 2938748,
"node_id": "MDQ6VXNlcjI5Mzg3NDg=",
"avatar_url": "https://avatars.githubusercontent.com/u/2938748?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mokkin",
"html_url": "https://github.com/mokkin",
"followers_url": "https://api.github.com/users/mokkin/followers",
"following_url": "https://api.github.com/users/mokkin/following{/other_user}",
"gists_url": "https://api.github.com/users/mokkin/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mokkin/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mokkin/subscriptions",
"organizations_url": "https://api.github.com/users/mokkin/orgs",
"repos_url": "https://api.github.com/users/mokkin/repos",
"events_url": "https://api.github.com/users/mokkin/events{/privacy}",
"received_events_url": "https://api.github.com/users/mokkin/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 35
| 2024-02-14T17:10:39
| 2024-08-05T22:20:08
| 2024-08-05T22:20:08
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I have tested Ollama on different machines yet, but no matter how many cores or RAM I have, it's only using 50% of the cores and just a very few GB of RAM.
For example now I'm running `ollama rum llama2:70b` on 16 core server with 32 GB of RAM, but while prompting only eight cores are used and just around 1 GB of RAM.
Is there something wrong? In the models descriptions are aleways warning you neet at least 8,16,32,... GB of RAM.

|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2496/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2496/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3754
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3754/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3754/comments
|
https://api.github.com/repos/ollama/ollama/issues/3754/events
|
https://github.com/ollama/ollama/issues/3754
| 2,253,202,923
|
I_kwDOJ0Z1Ps6GTSXr
| 3,754
|
ollama rm Support deletion through ID
|
{
"login": "eryajf",
"id": 33259379,
"node_id": "MDQ6VXNlcjMzMjU5Mzc5",
"avatar_url": "https://avatars.githubusercontent.com/u/33259379?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eryajf",
"html_url": "https://github.com/eryajf",
"followers_url": "https://api.github.com/users/eryajf/followers",
"following_url": "https://api.github.com/users/eryajf/following{/other_user}",
"gists_url": "https://api.github.com/users/eryajf/gists{/gist_id}",
"starred_url": "https://api.github.com/users/eryajf/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eryajf/subscriptions",
"organizations_url": "https://api.github.com/users/eryajf/orgs",
"repos_url": "https://api.github.com/users/eryajf/repos",
"events_url": "https://api.github.com/users/eryajf/events{/privacy}",
"received_events_url": "https://api.github.com/users/eryajf/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-04-19T14:48:40
| 2024-04-19T14:48:40
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Just like Docker RM supports deletion through ID, have you considered supporting model deletion through ID? If so, I am willing to research this area and submit a PR
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3754/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3754/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/225
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/225/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/225/comments
|
https://api.github.com/repos/ollama/ollama/issues/225/events
|
https://github.com/ollama/ollama/pull/225
| 1,824,949,943
|
PR_kwDOJ0Z1Ps5Wk5I4
| 225
|
add stop conditions
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-27T18:28:50
| 2023-07-28T00:20:57
| 2023-07-28T00:20:56
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/225",
"html_url": "https://github.com/ollama/ollama/pull/225",
"diff_url": "https://github.com/ollama/ollama/pull/225.diff",
"patch_url": "https://github.com/ollama/ollama/pull/225.patch",
"merged_at": "2023-07-28T00:20:56"
}
|
resolves #140
resolves #217
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/225/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/225/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1653
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1653/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1653/comments
|
https://api.github.com/repos/ollama/ollama/issues/1653/events
|
https://github.com/ollama/ollama/issues/1653
| 2,052,245,322
|
I_kwDOJ0Z1Ps56UsdK
| 1,653
|
shell autocompletion
|
{
"login": "teto",
"id": 886074,
"node_id": "MDQ6VXNlcjg4NjA3NA==",
"avatar_url": "https://avatars.githubusercontent.com/u/886074?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/teto",
"html_url": "https://github.com/teto",
"followers_url": "https://api.github.com/users/teto/followers",
"following_url": "https://api.github.com/users/teto/following{/other_user}",
"gists_url": "https://api.github.com/users/teto/gists{/gist_id}",
"starred_url": "https://api.github.com/users/teto/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/teto/subscriptions",
"organizations_url": "https://api.github.com/users/teto/orgs",
"repos_url": "https://api.github.com/users/teto/repos",
"events_url": "https://api.github.com/users/teto/events{/privacy}",
"received_events_url": "https://api.github.com/users/teto/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 6
| 2023-12-21T11:28:48
| 2024-12-04T05:19:01
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
hi, I would like to add shell autocompletion for ollama to my linux distribution.
I dont know go but the cli parser seems to be using https://github.com/spf13/cobra, the same as github CLI `gh`.
The nix package generates the autocompletion like this:
https://github.com/teto/nixpkgs/blob/dbaa028d61848fda534a04ce21459b722f0bca81/pkgs/applications/version-management/gh/default.nix#L30
so I tried `ollama completion -s zsh` hoping it was not shown by `ollama --help` because hidden but that failed.
Is there a way to generate those files currently and if yes, how, if not, could you add such a subcommand please ? makes writing commands on the prompt easier.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1653/reactions",
"total_count": 23,
"+1": 23,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1653/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/3365
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3365/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3365/comments
|
https://api.github.com/repos/ollama/ollama/issues/3365/events
|
https://github.com/ollama/ollama/issues/3365
| 2,209,739,153
|
I_kwDOJ0Z1Ps6DtfGR
| 3,365
|
Unable to make cors work in docker container
|
{
"login": "gbarton",
"id": 1878953,
"node_id": "MDQ6VXNlcjE4Nzg5NTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1878953?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gbarton",
"html_url": "https://github.com/gbarton",
"followers_url": "https://api.github.com/users/gbarton/followers",
"following_url": "https://api.github.com/users/gbarton/following{/other_user}",
"gists_url": "https://api.github.com/users/gbarton/gists{/gist_id}",
"starred_url": "https://api.github.com/users/gbarton/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gbarton/subscriptions",
"organizations_url": "https://api.github.com/users/gbarton/orgs",
"repos_url": "https://api.github.com/users/gbarton/repos",
"events_url": "https://api.github.com/users/gbarton/events{/privacy}",
"received_events_url": "https://api.github.com/users/gbarton/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 3
| 2024-03-27T03:24:08
| 2024-05-02T02:15:59
| 2024-03-27T22:44:09
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I have been trying to get cors to work inside a docker container with no luck so far. Attached is my docker-compose.yaml for starting ollama. I have tried many variations on the OLLAMA_HOST and OLLAMA_ORIGINS env vars to no luck. Does anyone know the magic combo to make it work?
### What did you expect to see?
hitting it from a browser directly works fine, from code works fine, using Ollama chats https://github.com/drazdra/ollama-chats hosted on another node runs into cors fun:

curling on the local box shows anything set as origin is blocked:
```bash
gman@ai:~/wrk/ai-runtime$ curl -i -H 'Origin:http://10.22.22.1' localhost:11434/
HTTP/1.1 403 Forbidden
Date: Wed, 27 Mar 2024 03:20:24 GMT
Content-Length: 0
gman@ai:~/wrk/ai-runtime$ curl localhost:11434
Ollama is running
```
### Steps to reproduce
docker-compose.yaml
```
version: '3.6'
services:
ollama:
container_name: ollama
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: 1
capabilities: [gpu]
image: ollama/ollama
#tty: true
restart: unless-stopped
environment:
- OLLAMA_ORIGINS="*"
- OLLAMA_HOST="0.0.0.0:11434"
ports:
- 11434:11434
volumes:
- ./ollama:/root/.ollama:cached
```
### Are there any recent changes that introduced the issue?
_No response_
### OS
Linux
### Architecture
amd64
### Platform
Docker
### Ollama version
0.1.29
### GPU
Nvidia
### GPU info
_No response_
### CPU
AMD
### Other software
_No response_
|
{
"login": "gbarton",
"id": 1878953,
"node_id": "MDQ6VXNlcjE4Nzg5NTM=",
"avatar_url": "https://avatars.githubusercontent.com/u/1878953?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gbarton",
"html_url": "https://github.com/gbarton",
"followers_url": "https://api.github.com/users/gbarton/followers",
"following_url": "https://api.github.com/users/gbarton/following{/other_user}",
"gists_url": "https://api.github.com/users/gbarton/gists{/gist_id}",
"starred_url": "https://api.github.com/users/gbarton/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gbarton/subscriptions",
"organizations_url": "https://api.github.com/users/gbarton/orgs",
"repos_url": "https://api.github.com/users/gbarton/repos",
"events_url": "https://api.github.com/users/gbarton/events{/privacy}",
"received_events_url": "https://api.github.com/users/gbarton/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3365/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3365/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2686
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2686/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2686/comments
|
https://api.github.com/repos/ollama/ollama/issues/2686/events
|
https://github.com/ollama/ollama/pull/2686
| 2,149,472,407
|
PR_kwDOJ0Z1Ps5nqz9L
| 2,686
|
update README to add Gemma 2B, 7B model in Model Library Table
|
{
"login": "adminazhar",
"id": 20738252,
"node_id": "MDQ6VXNlcjIwNzM4MjUy",
"avatar_url": "https://avatars.githubusercontent.com/u/20738252?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/adminazhar",
"html_url": "https://github.com/adminazhar",
"followers_url": "https://api.github.com/users/adminazhar/followers",
"following_url": "https://api.github.com/users/adminazhar/following{/other_user}",
"gists_url": "https://api.github.com/users/adminazhar/gists{/gist_id}",
"starred_url": "https://api.github.com/users/adminazhar/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/adminazhar/subscriptions",
"organizations_url": "https://api.github.com/users/adminazhar/orgs",
"repos_url": "https://api.github.com/users/adminazhar/repos",
"events_url": "https://api.github.com/users/adminazhar/events{/privacy}",
"received_events_url": "https://api.github.com/users/adminazhar/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-22T16:36:57
| 2024-02-22T18:07:47
| 2024-02-22T18:07:47
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2686",
"html_url": "https://github.com/ollama/ollama/pull/2686",
"diff_url": "https://github.com/ollama/ollama/pull/2686.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2686.patch",
"merged_at": "2024-02-22T18:07:47"
}
|
@pfrankov @jmorganca Kindly Review
|
{
"login": "mchiang0610",
"id": 3325447,
"node_id": "MDQ6VXNlcjMzMjU0NDc=",
"avatar_url": "https://avatars.githubusercontent.com/u/3325447?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mchiang0610",
"html_url": "https://github.com/mchiang0610",
"followers_url": "https://api.github.com/users/mchiang0610/followers",
"following_url": "https://api.github.com/users/mchiang0610/following{/other_user}",
"gists_url": "https://api.github.com/users/mchiang0610/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mchiang0610/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mchiang0610/subscriptions",
"organizations_url": "https://api.github.com/users/mchiang0610/orgs",
"repos_url": "https://api.github.com/users/mchiang0610/repos",
"events_url": "https://api.github.com/users/mchiang0610/events{/privacy}",
"received_events_url": "https://api.github.com/users/mchiang0610/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2686/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2686/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2911
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2911/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2911/comments
|
https://api.github.com/repos/ollama/ollama/issues/2911/events
|
https://github.com/ollama/ollama/issues/2911
| 2,166,737,391
|
I_kwDOJ0Z1Ps6BJcnv
| 2,911
|
System Monitor detected that internet is used by Ollama.
|
{
"login": "kukidevs",
"id": 113847173,
"node_id": "U_kgDOBskrhQ",
"avatar_url": "https://avatars.githubusercontent.com/u/113847173?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kukidevs",
"html_url": "https://github.com/kukidevs",
"followers_url": "https://api.github.com/users/kukidevs/followers",
"following_url": "https://api.github.com/users/kukidevs/following{/other_user}",
"gists_url": "https://api.github.com/users/kukidevs/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kukidevs/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kukidevs/subscriptions",
"organizations_url": "https://api.github.com/users/kukidevs/orgs",
"repos_url": "https://api.github.com/users/kukidevs/repos",
"events_url": "https://api.github.com/users/kukidevs/events{/privacy}",
"received_events_url": "https://api.github.com/users/kukidevs/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-03-04T12:29:53
| 2025-01-30T01:41:17
| 2024-03-04T17:50:06
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |

What does it send/receive? I was using it through MacOS terminal, installed the app from the website.
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2911/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2911/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2200
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2200/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2200/comments
|
https://api.github.com/repos/ollama/ollama/issues/2200/events
|
https://github.com/ollama/ollama/issues/2200
| 2,101,705,603
|
I_kwDOJ0Z1Ps59RXuD
| 2,200
|
ollama is stuck when translate language
|
{
"login": "MunhakKim",
"id": 13396311,
"node_id": "MDQ6VXNlcjEzMzk2MzEx",
"avatar_url": "https://avatars.githubusercontent.com/u/13396311?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MunhakKim",
"html_url": "https://github.com/MunhakKim",
"followers_url": "https://api.github.com/users/MunhakKim/followers",
"following_url": "https://api.github.com/users/MunhakKim/following{/other_user}",
"gists_url": "https://api.github.com/users/MunhakKim/gists{/gist_id}",
"starred_url": "https://api.github.com/users/MunhakKim/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MunhakKim/subscriptions",
"organizations_url": "https://api.github.com/users/MunhakKim/orgs",
"repos_url": "https://api.github.com/users/MunhakKim/repos",
"events_url": "https://api.github.com/users/MunhakKim/events{/privacy}",
"received_events_url": "https://api.github.com/users/MunhakKim/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-01-26T07:10:48
| 2024-03-11T23:14:33
| 2024-03-11T23:14:33
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
ollama is stuck when i ask to translate language.
In my case, i put it in like '한글로' for prompt which means 'in Korean' after English output.
Sometimes ollama could translate perfectly and stable, but mostly ollama is stuck.
I think if there is a problem in text generation, It should be pause not is broken.
In this case, **i have to restart ollama**.
I guess it is about string format from streaming.
i am using docker image from docker hub.
using this command 'docker exec -it ollama ollama run mistral'
or api.
version: v0.1.21 (i have been tested from v0.1.18~)
model: mistral, mixtral latest, llama2 etc.
sorry for no logs.
But it is easy to cause problems when you put '한글로' in prompt.
|
{
"login": "MunhakKim",
"id": 13396311,
"node_id": "MDQ6VXNlcjEzMzk2MzEx",
"avatar_url": "https://avatars.githubusercontent.com/u/13396311?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MunhakKim",
"html_url": "https://github.com/MunhakKim",
"followers_url": "https://api.github.com/users/MunhakKim/followers",
"following_url": "https://api.github.com/users/MunhakKim/following{/other_user}",
"gists_url": "https://api.github.com/users/MunhakKim/gists{/gist_id}",
"starred_url": "https://api.github.com/users/MunhakKim/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MunhakKim/subscriptions",
"organizations_url": "https://api.github.com/users/MunhakKim/orgs",
"repos_url": "https://api.github.com/users/MunhakKim/repos",
"events_url": "https://api.github.com/users/MunhakKim/events{/privacy}",
"received_events_url": "https://api.github.com/users/MunhakKim/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2200/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2200/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/870
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/870/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/870/comments
|
https://api.github.com/repos/ollama/ollama/issues/870/events
|
https://github.com/ollama/ollama/pull/870
| 1,955,433,858
|
PR_kwDOJ0Z1Ps5dcU35
| 870
|
Added a minimalist React UI for Ollama models to ReadME.md
|
{
"login": "richawo",
"id": 35015261,
"node_id": "MDQ6VXNlcjM1MDE1MjYx",
"avatar_url": "https://avatars.githubusercontent.com/u/35015261?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/richawo",
"html_url": "https://github.com/richawo",
"followers_url": "https://api.github.com/users/richawo/followers",
"following_url": "https://api.github.com/users/richawo/following{/other_user}",
"gists_url": "https://api.github.com/users/richawo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/richawo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/richawo/subscriptions",
"organizations_url": "https://api.github.com/users/richawo/orgs",
"repos_url": "https://api.github.com/users/richawo/repos",
"events_url": "https://api.github.com/users/richawo/events{/privacy}",
"received_events_url": "https://api.github.com/users/richawo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 4
| 2023-10-21T10:50:25
| 2023-10-23T14:44:39
| 2023-10-23T14:44:39
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/870",
"html_url": "https://github.com/ollama/ollama/pull/870",
"diff_url": "https://github.com/ollama/ollama/pull/870.diff",
"patch_url": "https://github.com/ollama/ollama/pull/870.patch",
"merged_at": "2023-10-23T14:44:39"
}
|
I discussed it a few times in the discord, and a few people seem to be using it, so it would be good to add.
Demo video:
https://github.com/jmorganca/ollama/assets/35015261/d50f7036-cdf2-44ed-9bb0-fdbed6a4ec66
I'll be maintaining/improving it a lot over the coming weeks, and some contributors reached out to get involved.
It can now handle markdown, etc. And continues to improve beyond the initial demo.
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/870/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/870/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2587
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2587/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2587/comments
|
https://api.github.com/repos/ollama/ollama/issues/2587/events
|
https://github.com/ollama/ollama/issues/2587
| 2,141,539,255
|
I_kwDOJ0Z1Ps5_pUu3
| 2,587
|
Running on GPU
|
{
"login": "shersoni610",
"id": 57876250,
"node_id": "MDQ6VXNlcjU3ODc2MjUw",
"avatar_url": "https://avatars.githubusercontent.com/u/57876250?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shersoni610",
"html_url": "https://github.com/shersoni610",
"followers_url": "https://api.github.com/users/shersoni610/followers",
"following_url": "https://api.github.com/users/shersoni610/following{/other_user}",
"gists_url": "https://api.github.com/users/shersoni610/gists{/gist_id}",
"starred_url": "https://api.github.com/users/shersoni610/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shersoni610/subscriptions",
"organizations_url": "https://api.github.com/users/shersoni610/orgs",
"repos_url": "https://api.github.com/users/shersoni610/repos",
"events_url": "https://api.github.com/users/shersoni610/events{/privacy}",
"received_events_url": "https://api.github.com/users/shersoni610/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 29
| 2024-02-19T05:27:39
| 2024-02-29T23:47:57
| 2024-02-29T23:47:57
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello,
It seems, the response time of llama2:7b is slow on my linux machine. I am not sure if the code
is running on Nvidia card.
In a python code, how to ensure that Ollama models run on GPU?
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2587/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2587/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5742
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5742/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5742/comments
|
https://api.github.com/repos/ollama/ollama/issues/5742/events
|
https://github.com/ollama/ollama/pull/5742
| 2,413,170,315
|
PR_kwDOJ0Z1Ps51nl1k
| 5,742
|
Added code to use swap memory in linux
|
{
"login": "avinash-ghadshi",
"id": 65444549,
"node_id": "MDQ6VXNlcjY1NDQ0NTQ5",
"avatar_url": "https://avatars.githubusercontent.com/u/65444549?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/avinash-ghadshi",
"html_url": "https://github.com/avinash-ghadshi",
"followers_url": "https://api.github.com/users/avinash-ghadshi/followers",
"following_url": "https://api.github.com/users/avinash-ghadshi/following{/other_user}",
"gists_url": "https://api.github.com/users/avinash-ghadshi/gists{/gist_id}",
"starred_url": "https://api.github.com/users/avinash-ghadshi/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/avinash-ghadshi/subscriptions",
"organizations_url": "https://api.github.com/users/avinash-ghadshi/orgs",
"repos_url": "https://api.github.com/users/avinash-ghadshi/repos",
"events_url": "https://api.github.com/users/avinash-ghadshi/events{/privacy}",
"received_events_url": "https://api.github.com/users/avinash-ghadshi/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 5
| 2024-07-17T09:54:10
| 2024-09-03T16:54:39
| 2024-09-03T16:54:39
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5742",
"html_url": "https://github.com/ollama/ollama/pull/5742",
"diff_url": "https://github.com/ollama/ollama/pull/5742.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5742.patch",
"merged_at": null
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5742/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5742/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6420
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6420/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6420/comments
|
https://api.github.com/repos/ollama/ollama/issues/6420/events
|
https://github.com/ollama/ollama/issues/6420
| 2,473,039,397
|
I_kwDOJ0Z1Ps6TZ5Yl
| 6,420
|
Is the speed of the Olama running model related to the CUDA version?
|
{
"login": "TianWuYuJiangHenShou",
"id": 20592000,
"node_id": "MDQ6VXNlcjIwNTkyMDAw",
"avatar_url": "https://avatars.githubusercontent.com/u/20592000?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/TianWuYuJiangHenShou",
"html_url": "https://github.com/TianWuYuJiangHenShou",
"followers_url": "https://api.github.com/users/TianWuYuJiangHenShou/followers",
"following_url": "https://api.github.com/users/TianWuYuJiangHenShou/following{/other_user}",
"gists_url": "https://api.github.com/users/TianWuYuJiangHenShou/gists{/gist_id}",
"starred_url": "https://api.github.com/users/TianWuYuJiangHenShou/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/TianWuYuJiangHenShou/subscriptions",
"organizations_url": "https://api.github.com/users/TianWuYuJiangHenShou/orgs",
"repos_url": "https://api.github.com/users/TianWuYuJiangHenShou/repos",
"events_url": "https://api.github.com/users/TianWuYuJiangHenShou/events{/privacy}",
"received_events_url": "https://api.github.com/users/TianWuYuJiangHenShou/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
},
{
"id": 6677745918,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g",
"url": "https://api.github.com/repos/ollama/ollama/labels/gpu",
"name": "gpu",
"color": "76C49E",
"default": false,
"description": ""
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3
| 2024-08-19T10:40:53
| 2024-09-26T00:13:40
| 2024-09-26T00:13:40
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I deployed qwen2:72B with the latest version of Ollama, but I found that the loading speed of Ollama models varies greatly under different nvidia driver version.
# driver:535.183.06 | cuda version:12.2
ollama version:0.3.4
Time of Loading Model:29s
# driver:515.105.01 | cuda version:11.7
ollama version:0.3.6
Time of Loading Model:659s
## GPU :A800
### OS
Linux
### GPU
Nvidia
### CPU
_No response_
### Ollama version
_No response_
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6420/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6420/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1418
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1418/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1418/comments
|
https://api.github.com/repos/ollama/ollama/issues/1418/events
|
https://github.com/ollama/ollama/issues/1418
| 2,031,151,545
|
I_kwDOJ0Z1Ps55EOm5
| 1,418
|
Make `notus` model available on `ollama`
|
{
"login": "davidberenstein1957",
"id": 25269220,
"node_id": "MDQ6VXNlcjI1MjY5MjIw",
"avatar_url": "https://avatars.githubusercontent.com/u/25269220?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/davidberenstein1957",
"html_url": "https://github.com/davidberenstein1957",
"followers_url": "https://api.github.com/users/davidberenstein1957/followers",
"following_url": "https://api.github.com/users/davidberenstein1957/following{/other_user}",
"gists_url": "https://api.github.com/users/davidberenstein1957/gists{/gist_id}",
"starred_url": "https://api.github.com/users/davidberenstein1957/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/davidberenstein1957/subscriptions",
"organizations_url": "https://api.github.com/users/davidberenstein1957/orgs",
"repos_url": "https://api.github.com/users/davidberenstein1957/repos",
"events_url": "https://api.github.com/users/davidberenstein1957/events{/privacy}",
"received_events_url": "https://api.github.com/users/davidberenstein1957/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 13
| 2023-12-07T16:53:09
| 2024-01-11T07:03:57
| 2024-01-11T07:03:57
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
We just hosted our [Argilla community meetup](https://www.youtube.com/playlist?list=PLBmuFBJ5cjcbsr49KFoC4DQoo3ZWT7q_d) and got the question if we would be able to make [Notus](https://argilla.io/blog/notus7b/) available on ollama. What would it take, and how can we help with this?
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1418/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1418/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1693
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1693/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1693/comments
|
https://api.github.com/repos/ollama/ollama/issues/1693/events
|
https://github.com/ollama/ollama/issues/1693
| 2,054,975,190
|
I_kwDOJ0Z1Ps56fG7W
| 1,693
|
Possible to increase speed / efficiency of model?
|
{
"login": "theyluvEnething",
"id": 115583828,
"node_id": "U_kgDOBuOrVA",
"avatar_url": "https://avatars.githubusercontent.com/u/115583828?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/theyluvEnething",
"html_url": "https://github.com/theyluvEnething",
"followers_url": "https://api.github.com/users/theyluvEnething/followers",
"following_url": "https://api.github.com/users/theyluvEnething/following{/other_user}",
"gists_url": "https://api.github.com/users/theyluvEnething/gists{/gist_id}",
"starred_url": "https://api.github.com/users/theyluvEnething/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/theyluvEnething/subscriptions",
"organizations_url": "https://api.github.com/users/theyluvEnething/orgs",
"repos_url": "https://api.github.com/users/theyluvEnething/repos",
"events_url": "https://api.github.com/users/theyluvEnething/events{/privacy}",
"received_events_url": "https://api.github.com/users/theyluvEnething/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 5
| 2023-12-24T03:49:05
| 2024-06-22T13:47:35
| 2024-03-12T00:16:40
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I'm trying out the Dolphin-Mixture model and it's quite fun, but really slow. (my specs are 64gb 3200mj ram, i7 4.5gh cpu and a 1080 ti)
but still it takes some time to start answering and when it starts it writes at maybe 2-3 words a second. Is it possible to make improvements to this?
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1693/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1693/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4172
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4172/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4172/comments
|
https://api.github.com/repos/ollama/ollama/issues/4172/events
|
https://github.com/ollama/ollama/pull/4172
| 2,279,565,700
|
PR_kwDOJ0Z1Ps5ukaTU
| 4,172
|
Add support for the BSD platforms
|
{
"login": "prep",
"id": 2017915,
"node_id": "MDQ6VXNlcjIwMTc5MTU=",
"avatar_url": "https://avatars.githubusercontent.com/u/2017915?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/prep",
"html_url": "https://github.com/prep",
"followers_url": "https://api.github.com/users/prep/followers",
"following_url": "https://api.github.com/users/prep/following{/other_user}",
"gists_url": "https://api.github.com/users/prep/gists{/gist_id}",
"starred_url": "https://api.github.com/users/prep/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/prep/subscriptions",
"organizations_url": "https://api.github.com/users/prep/orgs",
"repos_url": "https://api.github.com/users/prep/repos",
"events_url": "https://api.github.com/users/prep/events{/privacy}",
"received_events_url": "https://api.github.com/users/prep/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 6
| 2024-05-05T14:42:06
| 2024-09-08T12:35:59
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4172",
"html_url": "https://github.com/ollama/ollama/pull/4172",
"diff_url": "https://github.com/ollama/ollama/pull/4172.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4172.patch",
"merged_at": null
}
|
This PR adds the ability to compile and run Ollama on various BSD platforms, specifically DragonFly, FreeBSD, NetBSD and OpenBSD. A couple notes:
1. It includes support for Vulkan to offload work to the GPU.
2. It's only been tested on OpenBSD, but hopefully interested parties in this community can test the other platforms.
3. `ollama run` is broken on OpenBSD. The [readline](readline) package uses `syscall` which OpenBSD [has removed from libc and the kernel](https://marc.info/?l=openbsd-tech&m=169841790407370&w=2). Instead, the readline package should be updated to use `golang.org/x/sys/unix`, similar to how [ergochat/readline](https://github.com/ergochat/readline) does it. I didn't want to include that work in this PR.
4. The tweak in `gen_common.sh` is to remove a bash'ism so that the file can be sourced from `ksh`.
On OpenBSD I needed to add these packages:
```pkg_add cmake go vulkan-headers```
The other platforms will probably have similarly named packages to install.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4172/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4172/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8169
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8169/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8169/comments
|
https://api.github.com/repos/ollama/ollama/issues/8169/events
|
https://github.com/ollama/ollama/pull/8169
| 2,749,784,967
|
PR_kwDOJ0Z1Ps6FwPej
| 8,169
|
Update go.mod
|
{
"login": "Squishedmac",
"id": 88924339,
"node_id": "MDQ6VXNlcjg4OTI0MzM5",
"avatar_url": "https://avatars.githubusercontent.com/u/88924339?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Squishedmac",
"html_url": "https://github.com/Squishedmac",
"followers_url": "https://api.github.com/users/Squishedmac/followers",
"following_url": "https://api.github.com/users/Squishedmac/following{/other_user}",
"gists_url": "https://api.github.com/users/Squishedmac/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Squishedmac/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Squishedmac/subscriptions",
"organizations_url": "https://api.github.com/users/Squishedmac/orgs",
"repos_url": "https://api.github.com/users/Squishedmac/repos",
"events_url": "https://api.github.com/users/Squishedmac/events{/privacy}",
"received_events_url": "https://api.github.com/users/Squishedmac/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-12-19T09:50:26
| 2024-12-19T10:51:33
| 2024-12-19T10:51:33
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8169",
"html_url": "https://github.com/ollama/ollama/pull/8169",
"diff_url": "https://github.com/ollama/ollama/pull/8169.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8169.patch",
"merged_at": null
}
|
Fixed the CVE-2024-45337 severity on wiz scan for the docker image, changing the version of golang.org/x/crypto to the latest version should fix the issue
|
{
"login": "Squishedmac",
"id": 88924339,
"node_id": "MDQ6VXNlcjg4OTI0MzM5",
"avatar_url": "https://avatars.githubusercontent.com/u/88924339?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Squishedmac",
"html_url": "https://github.com/Squishedmac",
"followers_url": "https://api.github.com/users/Squishedmac/followers",
"following_url": "https://api.github.com/users/Squishedmac/following{/other_user}",
"gists_url": "https://api.github.com/users/Squishedmac/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Squishedmac/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Squishedmac/subscriptions",
"organizations_url": "https://api.github.com/users/Squishedmac/orgs",
"repos_url": "https://api.github.com/users/Squishedmac/repos",
"events_url": "https://api.github.com/users/Squishedmac/events{/privacy}",
"received_events_url": "https://api.github.com/users/Squishedmac/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8169/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8169/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6890
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6890/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6890/comments
|
https://api.github.com/repos/ollama/ollama/issues/6890/events
|
https://github.com/ollama/ollama/issues/6890
| 2,538,032,802
|
I_kwDOJ0Z1Ps6XR06i
| 6,890
|
"/show parameters" command causes crashes when running Qwen 2.5 models, on version 0.3.11
|
{
"login": "cmdntfnd",
"id": 142593649,
"node_id": "U_kgDOCH_OcQ",
"avatar_url": "https://avatars.githubusercontent.com/u/142593649?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/cmdntfnd",
"html_url": "https://github.com/cmdntfnd",
"followers_url": "https://api.github.com/users/cmdntfnd/followers",
"following_url": "https://api.github.com/users/cmdntfnd/following{/other_user}",
"gists_url": "https://api.github.com/users/cmdntfnd/gists{/gist_id}",
"starred_url": "https://api.github.com/users/cmdntfnd/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/cmdntfnd/subscriptions",
"organizations_url": "https://api.github.com/users/cmdntfnd/orgs",
"repos_url": "https://api.github.com/users/cmdntfnd/repos",
"events_url": "https://api.github.com/users/cmdntfnd/events{/privacy}",
"received_events_url": "https://api.github.com/users/cmdntfnd/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 1
| 2024-09-20T06:54:55
| 2024-12-13T17:07:10
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
This only happens after changing the parameters through /set parameter command.
Here's an example:
PS H:\ztmp> ollama run qwen2.5
>>> /show parameters
No parameters were specified for this model.
>>> /set parameter top_k 1
Set parameter 'top_k' to '1'
>>> /show parameters
error: couldn't get model
Error: something went wrong, please see the ollama server logs for details
PS H:\ztmp>
This is on windows 11.
Here's the error message from the ollama serve terminal tab:
2024/09/20 08:47:43 [Recovery] 2024/09/20 - 08:47:43 panic recovered:
assignment to entry in nil map
runtime/map_faststr.go:205 (0x7a93ba)
github.com/ollama/ollama/server/routes.go:807 (0x12cc57e)
github.com/ollama/ollama/server/routes.go:732 (0x12cb497)
github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1287cca)
github.com/ollama/ollama/server/routes.go:1076 (0x12d0c14)
github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1295d39)
github.com/gin-gonic/gin@v1.10.0/recovery.go:102 (0x1295d27)
github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1294e64)
github.com/gin-gonic/gin@v1.10.0/logger.go:249 (0x1294e4b)
github.com/gin-gonic/gin@v1.10.0/context.go:185 (0x1294291)
github.com/gin-gonic/gin@v1.10.0/gin.go:633 (0x1293d00)
github.com/gin-gonic/gin@v1.10.0/gin.go:589 (0x1293831)
net/http/server.go:2688 (0xafaecc)
net/http/server.go:3142 (0xafc6cd)
net/http/server.go:2044 (0xaf79c7)
runtime/asm_amd64.s:1695 (0x8026a0)
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.3.11
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6890/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6890/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/1020
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1020/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1020/comments
|
https://api.github.com/repos/ollama/ollama/issues/1020/events
|
https://github.com/ollama/ollama/pull/1020
| 1,979,730,196
|
PR_kwDOJ0Z1Ps5euED0
| 1,020
|
Add ModelFusion community integration
|
{
"login": "lgrammel",
"id": 205036,
"node_id": "MDQ6VXNlcjIwNTAzNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/205036?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lgrammel",
"html_url": "https://github.com/lgrammel",
"followers_url": "https://api.github.com/users/lgrammel/followers",
"following_url": "https://api.github.com/users/lgrammel/following{/other_user}",
"gists_url": "https://api.github.com/users/lgrammel/gists{/gist_id}",
"starred_url": "https://api.github.com/users/lgrammel/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lgrammel/subscriptions",
"organizations_url": "https://api.github.com/users/lgrammel/orgs",
"repos_url": "https://api.github.com/users/lgrammel/repos",
"events_url": "https://api.github.com/users/lgrammel/events{/privacy}",
"received_events_url": "https://api.github.com/users/lgrammel/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-11-06T17:54:56
| 2023-11-06T18:46:16
| 2023-11-06T18:46:16
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1020",
"html_url": "https://github.com/ollama/ollama/pull/1020",
"diff_url": "https://github.com/ollama/ollama/pull/1020.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1020.patch",
"merged_at": "2023-11-06T18:46:16"
}
| null |
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1020/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1020/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5656
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5656/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5656/comments
|
https://api.github.com/repos/ollama/ollama/issues/5656/events
|
https://github.com/ollama/ollama/pull/5656
| 2,406,435,963
|
PR_kwDOJ0Z1Ps51RF2g
| 5,656
|
llm: reorder gguf tensors
|
{
"login": "joshyan1",
"id": 76125168,
"node_id": "MDQ6VXNlcjc2MTI1MTY4",
"avatar_url": "https://avatars.githubusercontent.com/u/76125168?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/joshyan1",
"html_url": "https://github.com/joshyan1",
"followers_url": "https://api.github.com/users/joshyan1/followers",
"following_url": "https://api.github.com/users/joshyan1/following{/other_user}",
"gists_url": "https://api.github.com/users/joshyan1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/joshyan1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/joshyan1/subscriptions",
"organizations_url": "https://api.github.com/users/joshyan1/orgs",
"repos_url": "https://api.github.com/users/joshyan1/repos",
"events_url": "https://api.github.com/users/joshyan1/events{/privacy}",
"received_events_url": "https://api.github.com/users/joshyan1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-07-12T22:38:11
| 2024-08-14T16:49:46
| 2024-08-14T16:49:45
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5656",
"html_url": "https://github.com/ollama/ollama/pull/5656",
"diff_url": "https://github.com/ollama/ollama/pull/5656.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5656.patch",
"merged_at": null
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5656/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5656/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5629
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5629/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5629/comments
|
https://api.github.com/repos/ollama/ollama/issues/5629/events
|
https://github.com/ollama/ollama/issues/5629
| 2,403,190,147
|
I_kwDOJ0Z1Ps6PPcWD
| 5,629
|
Crashing or gibberish output on 3x Radeon GPUs
|
{
"login": "darwinvelez58",
"id": 118543481,
"node_id": "U_kgDOBxDUeQ",
"avatar_url": "https://avatars.githubusercontent.com/u/118543481?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/darwinvelez58",
"html_url": "https://github.com/darwinvelez58",
"followers_url": "https://api.github.com/users/darwinvelez58/followers",
"following_url": "https://api.github.com/users/darwinvelez58/following{/other_user}",
"gists_url": "https://api.github.com/users/darwinvelez58/gists{/gist_id}",
"starred_url": "https://api.github.com/users/darwinvelez58/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/darwinvelez58/subscriptions",
"organizations_url": "https://api.github.com/users/darwinvelez58/orgs",
"repos_url": "https://api.github.com/users/darwinvelez58/repos",
"events_url": "https://api.github.com/users/darwinvelez58/events{/privacy}",
"received_events_url": "https://api.github.com/users/darwinvelez58/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
}
] |
open
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 22
| 2024-07-11T13:26:43
| 2024-11-27T22:33:13
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
While running mixtral:8x7b-instruct-v0.1-q4_K_M on my physical machine with x3 7900 XTX I got this error:
```
[root@5dc6ecf27031 /]# ollama run mixtral:8x7b-instruct-v0.1-q4_K_M
Error: llama runner process has terminated: signal: segmentation fault (core dumped)
[root@5dc6ecf27031 /]#
```
Logs:
```
[GIN] 2024/07/11 - 13:22:44 | 200 | 16.23µs | 127.0.0.1 | HEAD "/"
[GIN] 2024/07/11 - 13:22:44 | 200 | 7.724554ms | 127.0.0.1 | POST "/api/show"
time=2024-07-11T13:22:44.297Z level=INFO source=sched.go:754 msg="new model will fit in available VRAM, loading" model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 library=rocm parallel=4 required="33.2 GiB"
time=2024-07-11T13:22:44.298Z level=INFO source=memory.go:309 msg="offload to rocm" layers.requested=-1 layers.model=33 layers.offload=33 layers.split=11,11,11 memory.available="[24.0 GiB 24.0 GiB 24.0 GiB]" memory.required.full="33.2 GiB" memory.required.partial="33.2 GiB" memory.required.kv="1.0 GiB" memory.required.allocations="[11.3 GiB 11.3 GiB 10.6 GiB]" memory.weights.total="25.5 GiB" memory.weights.repeating="25.4 GiB" memory.weights.nonrepeating="102.6 MiB" memory.graph.full="1.3 GiB" memory.graph.partial="1.3 GiB"
time=2024-07-11T13:22:44.299Z level=INFO source=server.go:375 msg="starting llama server" cmd="/tmp/ollama1419561683/runners/rocm_v60101/ollama_llama_server --model /root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 --ctx-size 8192 --batch-size 512 --embedding --log-disable --n-gpu-layers 33 --parallel 4 --tensor-split 11,11,11 --tensor-split 11,11,11 --port 41695"
time=2024-07-11T13:22:44.299Z level=INFO source=sched.go:474 msg="loaded runners" count=1
time=2024-07-11T13:22:44.299Z level=INFO source=server.go:563 msg="waiting for llama runner to start responding"
time=2024-07-11T13:22:44.299Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server error"
INFO [main] build info | build=1 commit="a8db2a9" tid="140134008951616" timestamp=1720704164
INFO [main] system info | n_threads=16 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 0 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 0 | " tid="140134008951616" timestamp=1720704164 total_threads=32
INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="31" port="41695" tid="140134008951616" timestamp=1720704164
llama_model_loader: loaded meta data with 26 key-value pairs and 995 tensors from /root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv 0: general.architecture str = llama
llama_model_loader: - kv 1: general.name str = mistralai
llama_model_loader: - kv 2: llama.context_length u32 = 32768
llama_model_loader: - kv 3: llama.embedding_length u32 = 4096
llama_model_loader: - kv 4: llama.block_count u32 = 32
llama_model_loader: - kv 5: llama.feed_forward_length u32 = 14336
llama_model_loader: - kv 6: llama.rope.dimension_count u32 = 128
llama_model_loader: - kv 7: llama.attention.head_count u32 = 32
llama_model_loader: - kv 8: llama.attention.head_count_kv u32 = 8
llama_model_loader: - kv 9: llama.expert_count u32 = 8
llama_model_loader: - kv 10: llama.expert_used_count u32 = 2
llama_model_loader: - kv 11: llama.attention.layer_norm_rms_epsilon f32 = 0.000010
llama_model_loader: - kv 12: llama.rope.freq_base f32 = 1000000.000000
llama_model_loader: - kv 13: general.file_type u32 = 15
llama_model_loader: - kv 14: tokenizer.ggml.model str = llama
llama_model_loader: - kv 15: tokenizer.ggml.tokens arr[str,32000] = ["<unk>", "<s>", "</s>", "<0x00>", "<...
llama_model_loader: - kv 16: tokenizer.ggml.scores arr[f32,32000] = [0.000000, 0.000000, 0.000000, 0.0000...
llama_model_loader: - kv 17: tokenizer.ggml.token_type arr[i32,32000] = [2, 3, 3, 6, 6, 6, 6, 6, 6, 6, 6, 6, ...
llama_model_loader: - kv 18: tokenizer.ggml.merges arr[str,58980] = ["▁ t", "i n", "e r", "▁ a", "h e...
llama_model_loader: - kv 19: tokenizer.ggml.bos_token_id u32 = 1
llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 2
llama_model_loader: - kv 21: tokenizer.ggml.unknown_token_id u32 = 0
llama_model_loader: - kv 22: tokenizer.ggml.add_bos_token bool = true
llama_model_loader: - kv 23: tokenizer.ggml.add_eos_token bool = false
llama_model_loader: - kv 24: tokenizer.chat_template str = {{ bos_token }}{% for message in mess...
llama_model_loader: - kv 25: general.quantization_version u32 = 2
llama_model_loader: - type f32: 65 tensors
llama_model_loader: - type f16: 32 tensors
llama_model_loader: - type q8_0: 64 tensors
llama_model_loader: - type q4_K: 833 tensors
llama_model_loader: - type q6_K: 1 tensors
llm_load_vocab: special tokens cache size = 259
llm_load_vocab: token to piece cache size = 0.1637 MB
llm_load_print_meta: format = GGUF V3 (latest)
llm_load_print_meta: arch = llama
llm_load_print_meta: vocab type = SPM
llm_load_print_meta: n_vocab = 32000
llm_load_print_meta: n_merges = 0
llm_load_print_meta: vocab_only = 0
llm_load_print_meta: n_ctx_train = 32768
llm_load_print_meta: n_embd = 4096
llm_load_print_meta: n_layer = 32
llm_load_print_meta: n_head = 32
llm_load_print_meta: n_head_kv = 8
llm_load_print_meta: n_rot = 128
llm_load_print_meta: n_swa = 0
llm_load_print_meta: n_embd_head_k = 128
llm_load_print_meta: n_embd_head_v = 128
llm_load_print_meta: n_gqa = 4
llm_load_print_meta: n_embd_k_gqa = 1024
llm_load_print_meta: n_embd_v_gqa = 1024
llm_load_print_meta: f_norm_eps = 0.0e+00
llm_load_print_meta: f_norm_rms_eps = 1.0e-05
llm_load_print_meta: f_clamp_kqv = 0.0e+00
llm_load_print_meta: f_max_alibi_bias = 0.0e+00
llm_load_print_meta: f_logit_scale = 0.0e+00
llm_load_print_meta: n_ff = 14336
llm_load_print_meta: n_expert = 8
llm_load_print_meta: n_expert_used = 2
llm_load_print_meta: causal attn = 1
llm_load_print_meta: pooling type = 0
llm_load_print_meta: rope type = 0
llm_load_print_meta: rope scaling = linear
llm_load_print_meta: freq_base_train = 1000000.0
llm_load_print_meta: freq_scale_train = 1
llm_load_print_meta: n_ctx_orig_yarn = 32768
llm_load_print_meta: rope_finetuned = unknown
llm_load_print_meta: ssm_d_conv = 0
llm_load_print_meta: ssm_d_inner = 0
llm_load_print_meta: ssm_d_state = 0
llm_load_print_meta: ssm_dt_rank = 0
llm_load_print_meta: model type = 8x7B
llm_load_print_meta: model ftype = Q4_K - Medium
llm_load_print_meta: model params = 46.70 B
llm_load_print_meta: model size = 24.62 GiB (4.53 BPW)
llm_load_print_meta: general.name = mistralai
llm_load_print_meta: BOS token = 1 '<s>'
llm_load_print_meta: EOS token = 2 '</s>'
llm_load_print_meta: UNK token = 0 '<unk>'
llm_load_print_meta: LF token = 13 '<0x0A>'
llm_load_print_meta: max token length = 48
time=2024-07-11T13:22:44.549Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server loading model"
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 ROCm devices:
Device 0: Radeon RX 7900 XTX, compute capability 11.0, VMM: no
Device 1: Radeon RX 7900 XTX, compute capability 11.0, VMM: no
Device 2: Radeon RX 7900 XTX, compute capability 11.0, VMM: no
llm_load_tensors: ggml ctx size = 1.53 MiB
llm_load_tensors: offloading 32 repeating layers to GPU
llm_load_tensors: offloading non-repeating layers to GPU
llm_load_tensors: offloaded 33/33 layers to GPU
llm_load_tensors: ROCm0 buffer size = 8608.53 MiB
llm_load_tensors: ROCm1 buffer size = 8608.53 MiB
llm_load_tensors: ROCm2 buffer size = 7928.49 MiB
llm_load_tensors: ROCm_Host buffer size = 70.31 MiB
time=2024-07-11T13:23:03.566Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server not responding"
llama_new_context_with_model: n_ctx = 8192
llama_new_context_with_model: n_batch = 512
llama_new_context_with_model: n_ubatch = 512
llama_new_context_with_model: flash_attn = 0
llama_new_context_with_model: freq_base = 1000000.0
llama_new_context_with_model: freq_scale = 1
time=2024-07-11T13:23:04.460Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server loading model"
llama_kv_cache_init: ROCm0 KV buffer size = 352.00 MiB
llama_kv_cache_init: ROCm1 KV buffer size = 352.00 MiB
llama_kv_cache_init: ROCm2 KV buffer size = 320.00 MiB
llama_new_context_with_model: KV self size = 1024.00 MiB, K (f16): 512.00 MiB, V (f16): 512.00 MiB
llama_new_context_with_model: ROCm_Host output buffer size = 0.55 MiB
llama_new_context_with_model: pipeline parallelism enabled (n_copies=4)
llama_new_context_with_model: ROCm0 compute buffer size = 640.01 MiB
llama_new_context_with_model: ROCm1 compute buffer size = 640.01 MiB
llama_new_context_with_model: ROCm2 compute buffer size = 640.02 MiB
llama_new_context_with_model: ROCm_Host compute buffer size = 72.02 MiB
llama_new_context_with_model: graph nodes = 1510
llama_new_context_with_model: graph splits = 4
time=2024-07-11T13:23:06.864Z level=INFO source=server.go:604 msg="waiting for server to become available" status="llm server error"
[GIN] 2024/07/11 - 13:23:07 | 500 | 22.834580361s | 127.0.0.1 | POST "/api/chat"
time=2024-07-11T13:23:07.115Z level=ERROR source=sched.go:480 msg="error loading llama server" error="llama runner process has terminated: signal: segmentation fault (core dumped) "
time=2024-07-11T13:23:12.116Z level=WARN source=sched.go:671 msg="gpu VRAM usage didn't recover within timeout" seconds=5.001085328 model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8
time=2024-07-11T13:23:12.366Z level=WARN source=sched.go:671 msg="gpu VRAM usage didn't recover within timeout" seconds=5.251122065 model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8
time=2024-07-11T13:23:12.616Z level=WARN source=sched.go:671 msg="gpu VRAM usage didn't recover within timeout" seconds=5.500799906 model=/root/.ollama/models/blobs/sha256-3a17f7cde150070bbc815645693fb93c311cc42e7deaf198364acadcf08458f8
```
I am running this docker version
docker run -d --restart unless-stopped --device /dev/kfd --device /dev/dri -v ollama:/root/.ollama -p 11442:11434 --name dvz3 ollama/ollama:0.2.1-rocm
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.2.1-rocm
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5629/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5629/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/1707
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1707/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1707/comments
|
https://api.github.com/repos/ollama/ollama/issues/1707/events
|
https://github.com/ollama/ollama/issues/1707
| 2,055,634,785
|
I_kwDOJ0Z1Ps56hn9h
| 1,707
|
[Bug] Calling embedding endpoint within short time will receive `None`
|
{
"login": "samx81",
"id": 20315145,
"node_id": "MDQ6VXNlcjIwMzE1MTQ1",
"avatar_url": "https://avatars.githubusercontent.com/u/20315145?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/samx81",
"html_url": "https://github.com/samx81",
"followers_url": "https://api.github.com/users/samx81/followers",
"following_url": "https://api.github.com/users/samx81/following{/other_user}",
"gists_url": "https://api.github.com/users/samx81/gists{/gist_id}",
"starred_url": "https://api.github.com/users/samx81/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/samx81/subscriptions",
"organizations_url": "https://api.github.com/users/samx81/orgs",
"repos_url": "https://api.github.com/users/samx81/repos",
"events_url": "https://api.github.com/users/samx81/events{/privacy}",
"received_events_url": "https://api.github.com/users/samx81/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 1
| 2023-12-25T11:44:12
| 2024-05-06T23:44:40
| 2024-05-06T23:44:40
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I'm currently using `llama_index` to make LLM do a document QA,
but I notice the embedding endpoint often output:
```json
{"embedding":null}
```
when I do `condense_question` chat mode which calls LLM to rephrase question (and match context) before actually answer.
Like:
```python
llm = Ollama(model="dolphin2.2-mistral:7b-q4_K_M")
ollama_embedding = OllamaEmbedding(
model_name="dolphin2.2-mistral:7b-q4_K_M",
ollama_additional_kwargs={"mirostat": 0},
)
service_context = ServiceContext.from_defaults(llm=llm, embed_model=ollama_embedding)
chat_engine = index.as_chat_engine(service_context=service_context,chat_mode='condense_plus_context')
resp= chat_engine.chat('abc')
print(resp)
# Often happens when LLM is called second time
resp= chat_engine.chat('cde')
print(resp)
```
I guess it is probably due to ollama or llama.cpp needs to handle different behavior in the same time?
Because if I modify `llama_index` to check for `null` and retry, the above code will work:
```python
## In llama_index.embeddings.OllamaEmbedding
while True:
response = requests.post(
url=f"{self.base_url}/api/embeddings",
headers={"Content-Type": "application/json"},
json=ollama_request_body,
)
response.encoding = "utf-8"
if response.status_code != 200:
optional_detail = response.json().get("error")
raise ValueError(
f"Ollama call failed with status code {response.status_code}."
f" Details: {optional_detail}"
)
if response.json()["embedding"]:
break
```
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1707/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1707/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7605
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7605/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7605/comments
|
https://api.github.com/repos/ollama/ollama/issues/7605/events
|
https://github.com/ollama/ollama/issues/7605
| 2,647,941,751
|
I_kwDOJ0Z1Ps6d1GJ3
| 7,605
|
[Model request]GOT-OCR2.0
|
{
"login": "monkeycc",
"id": 6490927,
"node_id": "MDQ6VXNlcjY0OTA5Mjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/6490927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/monkeycc",
"html_url": "https://github.com/monkeycc",
"followers_url": "https://api.github.com/users/monkeycc/followers",
"following_url": "https://api.github.com/users/monkeycc/following{/other_user}",
"gists_url": "https://api.github.com/users/monkeycc/gists{/gist_id}",
"starred_url": "https://api.github.com/users/monkeycc/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/monkeycc/subscriptions",
"organizations_url": "https://api.github.com/users/monkeycc/orgs",
"repos_url": "https://api.github.com/users/monkeycc/repos",
"events_url": "https://api.github.com/users/monkeycc/events{/privacy}",
"received_events_url": "https://api.github.com/users/monkeycc/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
open
| false
| null |
[] | null | 0
| 2024-11-11T02:30:06
| 2024-12-18T01:46:27
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
https://github.com/Ucas-HaoranWei/GOT-OCR2.0
|
{
"login": "monkeycc",
"id": 6490927,
"node_id": "MDQ6VXNlcjY0OTA5Mjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/6490927?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/monkeycc",
"html_url": "https://github.com/monkeycc",
"followers_url": "https://api.github.com/users/monkeycc/followers",
"following_url": "https://api.github.com/users/monkeycc/following{/other_user}",
"gists_url": "https://api.github.com/users/monkeycc/gists{/gist_id}",
"starred_url": "https://api.github.com/users/monkeycc/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/monkeycc/subscriptions",
"organizations_url": "https://api.github.com/users/monkeycc/orgs",
"repos_url": "https://api.github.com/users/monkeycc/repos",
"events_url": "https://api.github.com/users/monkeycc/events{/privacy}",
"received_events_url": "https://api.github.com/users/monkeycc/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7605/reactions",
"total_count": 6,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 3,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7605/timeline
| null |
reopened
| false
|
https://api.github.com/repos/ollama/ollama/issues/7774
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7774/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7774/comments
|
https://api.github.com/repos/ollama/ollama/issues/7774/events
|
https://github.com/ollama/ollama/issues/7774
| 2,677,843,192
|
I_kwDOJ0Z1Ps6fnKT4
| 7,774
|
Toolcall stream
|
{
"login": "ZHOUxiaohe1987",
"id": 59469405,
"node_id": "MDQ6VXNlcjU5NDY5NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/59469405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ZHOUxiaohe1987",
"html_url": "https://github.com/ZHOUxiaohe1987",
"followers_url": "https://api.github.com/users/ZHOUxiaohe1987/followers",
"following_url": "https://api.github.com/users/ZHOUxiaohe1987/following{/other_user}",
"gists_url": "https://api.github.com/users/ZHOUxiaohe1987/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ZHOUxiaohe1987/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ZHOUxiaohe1987/subscriptions",
"organizations_url": "https://api.github.com/users/ZHOUxiaohe1987/orgs",
"repos_url": "https://api.github.com/users/ZHOUxiaohe1987/repos",
"events_url": "https://api.github.com/users/ZHOUxiaohe1987/events{/privacy}",
"received_events_url": "https://api.github.com/users/ZHOUxiaohe1987/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 0
| 2024-11-21T03:21:31
| 2024-11-21T03:22:11
| 2024-11-21T03:21:38
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null | null |
{
"login": "ZHOUxiaohe1987",
"id": 59469405,
"node_id": "MDQ6VXNlcjU5NDY5NDA1",
"avatar_url": "https://avatars.githubusercontent.com/u/59469405?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ZHOUxiaohe1987",
"html_url": "https://github.com/ZHOUxiaohe1987",
"followers_url": "https://api.github.com/users/ZHOUxiaohe1987/followers",
"following_url": "https://api.github.com/users/ZHOUxiaohe1987/following{/other_user}",
"gists_url": "https://api.github.com/users/ZHOUxiaohe1987/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ZHOUxiaohe1987/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ZHOUxiaohe1987/subscriptions",
"organizations_url": "https://api.github.com/users/ZHOUxiaohe1987/orgs",
"repos_url": "https://api.github.com/users/ZHOUxiaohe1987/repos",
"events_url": "https://api.github.com/users/ZHOUxiaohe1987/events{/privacy}",
"received_events_url": "https://api.github.com/users/ZHOUxiaohe1987/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7774/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7774/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8597
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8597/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8597/comments
|
https://api.github.com/repos/ollama/ollama/issues/8597/events
|
https://github.com/ollama/ollama/issues/8597
| 2,811,701,408
|
I_kwDOJ0Z1Ps6nlyig
| 8,597
|
Error: llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer (4x L40S, 384GB system RAM, Deepseek-R1)
|
{
"login": "orlyandico",
"id": 1325420,
"node_id": "MDQ6VXNlcjEzMjU0MjA=",
"avatar_url": "https://avatars.githubusercontent.com/u/1325420?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/orlyandico",
"html_url": "https://github.com/orlyandico",
"followers_url": "https://api.github.com/users/orlyandico/followers",
"following_url": "https://api.github.com/users/orlyandico/following{/other_user}",
"gists_url": "https://api.github.com/users/orlyandico/gists{/gist_id}",
"starred_url": "https://api.github.com/users/orlyandico/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/orlyandico/subscriptions",
"organizations_url": "https://api.github.com/users/orlyandico/orgs",
"repos_url": "https://api.github.com/users/orlyandico/repos",
"events_url": "https://api.github.com/users/orlyandico/events{/privacy}",
"received_events_url": "https://api.github.com/users/orlyandico/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 17
| 2025-01-26T17:56:11
| 2025-01-30T07:47:39
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I am aware that 4 x L40S only has 192GB of VRAM, but the machine also has 384GB of system RAM. The error seems to indicate that 45108 MiB of RAM is being allocated with cudaMalloc and this is failing. This is very close to the GPU limit (46068 MiB). On my home setup (2x P40, admittedly not trying such a huge model) the GPU never gets close to its VRAM limit (typically only 22GB out of 24GB).
```
Jan 26 17:48:20 ip-172-31-3-6 ollama[1418]: llm_load_print_meta: rope_yarn_log_mul = 0.1000
Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: ggml_backend_cuda_buffer_type_alloc_buffer: allocating 45108.64 MiB on device 0: cudaMalloc failed: out of memory
Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: llama_model_load: error loading model: unable to allocate CUDA0 buffer
Jan 26 17:50:35 ip-172-31-3-6 ollama[1418]: llama_load_model_from_file: failed to load model
Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: panic: unable to load model: /usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9
Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: goroutine 34 [running]:
Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner.(*Server).loadModel(0xc0001a0000, {0x1b, 0x0, 0x0, 0x0, {0xc000194090, 0x4, 0x4}, 0xc00018a060, 0x0}, ...)
Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner/runner.go:852 +0x3ad
Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: created by github.com/ollama/ollama/llama/runner.Execute in goroutine 1
Jan 26 17:51:23 ip-172-31-3-6 ollama[1418]: github.com/ollama/ollama/llama/runner/runner.go:970 +0xd0d
Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:24.077Z level=INFO source=server.go:589 msg="waiting for server to become available" status="llm server error"
Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:24.578Z level=ERROR source=sched.go:455 msg="error loading llama server" error="llama runner process has terminated: error loading model: unable to allocate CUDA0 buffer\nllama_load_model_from_file: failed to load model"
Jan 26 17:51:24 ip-172-31-3-6 ollama[1418]: [GIN] 2025/01/26 - 17:51:24 | 500 | 3m6s | 127.0.0.1 | POST "/api/generate"
Jan 26 17:51:30 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:30.190Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=5.6116838510000004 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1>
Jan 26 17:51:30 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:30.947Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=6.369278328 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9
Jan 26 17:51:31 ip-172-31-3-6 ollama[1418]: time=2025-01-26T17:51:31.707Z level=WARN source=sched.go:646 msg="gpu VRAM usage didn't recover within timeout" seconds=7.128851916 model=/usr/share/ollama/.ollama/models/blobs/sha256-9801e7fce27dbf3d0bfb468b7b21f1d132131a546dfc43e50518631b8b1800a9
Jan 26 17:51:40 ip-172-31-3-6 ollama[1418]: [GIN] 2025/01/26 - 17:51:40 | 200 | 65.972µs | 127.0.0.1 | GET "/api/version"
ubuntu@ip-172-31-3-6:~$ ollama --version
ollama version is 0.5.7
ubuntu@ip-172-31-3-6:~$ free
total used free shared buff/cache available
Mem: 390837004 4152012 385439624 3248 1245368 383763992
Swap: 0 0 0
ubuntu@ip-172-31-3-6:~$ nvidia-smi
Sun Jan 26 17:48:08 2025
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.144.03 Driver Version: 550.144.03 CUDA Version: 12.4 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA L40S On | 00000000:38:00.0 Off | 0 |
| N/A 22C P8 22W / 350W | 4MiB / 46068MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 1 NVIDIA L40S On | 00000000:3A:00.0 Off | 0 |
| N/A 22C P8 21W / 350W | 4MiB / 46068MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 2 NVIDIA L40S On | 00000000:3C:00.0 Off | 0 |
| N/A 22C P8 44W / 350W | 4MiB / 46068MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 3 NVIDIA L40S On | 00000000:3E:00.0 Off | 0 |
| N/A 23C P8 22W / 350W | 4MiB / 46068MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
```
### OS
Linux
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.5.7
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8597/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8597/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/7270
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7270/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7270/comments
|
https://api.github.com/repos/ollama/ollama/issues/7270/events
|
https://github.com/ollama/ollama/pull/7270
| 2,599,159,642
|
PR_kwDOJ0Z1Ps5_LPMN
| 7,270
|
Show supported types list in help message for `ollama create`
|
{
"login": "shirayu",
"id": 963961,
"node_id": "MDQ6VXNlcjk2Mzk2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/963961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shirayu",
"html_url": "https://github.com/shirayu",
"followers_url": "https://api.github.com/users/shirayu/followers",
"following_url": "https://api.github.com/users/shirayu/following{/other_user}",
"gists_url": "https://api.github.com/users/shirayu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/shirayu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shirayu/subscriptions",
"organizations_url": "https://api.github.com/users/shirayu/orgs",
"repos_url": "https://api.github.com/users/shirayu/repos",
"events_url": "https://api.github.com/users/shirayu/events{/privacy}",
"received_events_url": "https://api.github.com/users/shirayu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-10-19T13:15:48
| 2024-11-27T06:43:00
| 2024-11-27T06:43:00
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7270",
"html_url": "https://github.com/ollama/ollama/pull/7270",
"diff_url": "https://github.com/ollama/ollama/pull/7270.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7270.patch",
"merged_at": null
}
|
This PR adds supported types list to the help message for `ollama create`
```console
$ go run main.go create -h
Create a model from a Modelfile
Usage:
ollama create MODEL [flags]
Flags:
-f, --file string Name of the Modelfile (default "Modelfile")
-h, --help help for create
-q, --quantize string Quantize model to this level (e.g. q4_0)
Supported types:
BF16, F16, F32, IQ1_M, IQ1_S, IQ2_M
IQ2_S, IQ2_XS, IQ2_XXS, IQ3_S, IQ3_XS, IQ3_XXS
IQ4_NL, IQ4_XS, Q2_K, Q2_K_S, Q3_K_L, Q3_K_M
Q3_K_S, Q4_0, Q4_1, Q4_1_F16, Q4_K_M, Q4_K_S
Q5_0, Q5_1, Q5_K_M, Q5_K_S, Q6_K, Q8_0
Environment Variables:
OLLAMA_HOST IP Address for the ollama server (default 127.0.0.1:11434)
```
|
{
"login": "shirayu",
"id": 963961,
"node_id": "MDQ6VXNlcjk2Mzk2MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/963961?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/shirayu",
"html_url": "https://github.com/shirayu",
"followers_url": "https://api.github.com/users/shirayu/followers",
"following_url": "https://api.github.com/users/shirayu/following{/other_user}",
"gists_url": "https://api.github.com/users/shirayu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/shirayu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/shirayu/subscriptions",
"organizations_url": "https://api.github.com/users/shirayu/orgs",
"repos_url": "https://api.github.com/users/shirayu/repos",
"events_url": "https://api.github.com/users/shirayu/events{/privacy}",
"received_events_url": "https://api.github.com/users/shirayu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7270/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7270/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/478
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/478/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/478/comments
|
https://api.github.com/repos/ollama/ollama/issues/478/events
|
https://github.com/ollama/ollama/pull/478
| 1,884,825,856
|
PR_kwDOJ0Z1Ps5ZufRs
| 478
|
remove unused openssh key types
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2023-09-06T22:13:05
| 2023-09-06T22:18:55
| 2023-09-06T22:18:54
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/478",
"html_url": "https://github.com/ollama/ollama/pull/478",
"diff_url": "https://github.com/ollama/ollama/pull/478.diff",
"patch_url": "https://github.com/ollama/ollama/pull/478.patch",
"merged_at": "2023-09-06T22:18:54"
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/478/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/478/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7369
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7369/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7369/comments
|
https://api.github.com/repos/ollama/ollama/issues/7369/events
|
https://github.com/ollama/ollama/pull/7369
| 2,615,356,549
|
PR_kwDOJ0Z1Ps5_94Uy
| 7,369
|
Fix deepseek deseret regex
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-10-25T23:37:50
| 2024-10-26T21:58:57
| 2024-10-26T21:58:54
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7369",
"html_url": "https://github.com/ollama/ollama/pull/7369",
"diff_url": "https://github.com/ollama/ollama/pull/7369.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7369.patch",
"merged_at": "2024-10-26T21:58:54"
}
|
On windows compiled with gcc the c++ regex library failed to handle the characters
Without any changes, loading the model in the Go server crashes with
```
llama_model_load: error loading model: error loading model vocabulary: wstring_convert::from_bytes
```
The patch for unicode.cpp gets past the wide character conversion problem, but then hits
```
Regex error: Invalid range in bracket expression.
```
Switching to the `\U<8hexchars>` syntax to capture the range for Deseret resolves the regex problem.
Fixes #7311
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7369/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7369/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4416
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4416/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4416/comments
|
https://api.github.com/repos/ollama/ollama/issues/4416/events
|
https://github.com/ollama/ollama/pull/4416
| 2,294,175,355
|
PR_kwDOJ0Z1Ps5vVSj2
| 4,416
|
don't abort when an invalid model name is used in /save
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-05-14T01:42:17
| 2024-05-14T01:48:29
| 2024-05-14T01:48:28
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4416",
"html_url": "https://github.com/ollama/ollama/pull/4416",
"diff_url": "https://github.com/ollama/ollama/pull/4416.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4416.patch",
"merged_at": "2024-05-14T01:48:28"
}
|
Fixes #3852
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4416/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4416/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2774
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2774/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2774/comments
|
https://api.github.com/repos/ollama/ollama/issues/2774/events
|
https://github.com/ollama/ollama/issues/2774
| 2,155,807,968
|
I_kwDOJ0Z1Ps6AfwTg
| 2,774
|
What is the different between /api/generate and /api/chat?
|
{
"login": "owenzhao",
"id": 2182896,
"node_id": "MDQ6VXNlcjIxODI4OTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/2182896?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/owenzhao",
"html_url": "https://github.com/owenzhao",
"followers_url": "https://api.github.com/users/owenzhao/followers",
"following_url": "https://api.github.com/users/owenzhao/following{/other_user}",
"gists_url": "https://api.github.com/users/owenzhao/gists{/gist_id}",
"starred_url": "https://api.github.com/users/owenzhao/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/owenzhao/subscriptions",
"organizations_url": "https://api.github.com/users/owenzhao/orgs",
"repos_url": "https://api.github.com/users/owenzhao/repos",
"events_url": "https://api.github.com/users/owenzhao/events{/privacy}",
"received_events_url": "https://api.github.com/users/owenzhao/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 16
| 2024-02-27T06:53:37
| 2024-09-09T18:39:13
| 2024-03-12T06:33:16
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I mean if I give them the same prompt and input, the answers will be the same. Right? Then why they are two different API?
Or is chat auto context? I mean when using /api/chat, the answer will automatically include the previous conversation? And the /api/generate only answer for the present?
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2774/reactions",
"total_count": 14,
"+1": 14,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2774/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8232
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8232/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8232/comments
|
https://api.github.com/repos/ollama/ollama/issues/8232/events
|
https://github.com/ollama/ollama/pull/8232
| 2,757,927,665
|
PR_kwDOJ0Z1Ps6GLO5h
| 8,232
|
Remove redundant image_url format check in fromChatRequest
|
{
"login": "muzzlol",
"id": 116030715,
"node_id": "U_kgDOBup8-w",
"avatar_url": "https://avatars.githubusercontent.com/u/116030715?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/muzzlol",
"html_url": "https://github.com/muzzlol",
"followers_url": "https://api.github.com/users/muzzlol/followers",
"following_url": "https://api.github.com/users/muzzlol/following{/other_user}",
"gists_url": "https://api.github.com/users/muzzlol/gists{/gist_id}",
"starred_url": "https://api.github.com/users/muzzlol/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/muzzlol/subscriptions",
"organizations_url": "https://api.github.com/users/muzzlol/orgs",
"repos_url": "https://api.github.com/users/muzzlol/repos",
"events_url": "https://api.github.com/users/muzzlol/events{/privacy}",
"received_events_url": "https://api.github.com/users/muzzlol/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1
| 2024-12-24T15:22:37
| 2025-01-13T19:32:37
| 2025-01-13T19:32:37
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8232",
"html_url": "https://github.com/ollama/ollama/pull/8232",
"diff_url": "https://github.com/ollama/ollama/pull/8232.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8232.patch",
"merged_at": null
}
|
Improves the `fromChatRequest` function by removing redundant checks for Ollama's direct string format and improving error messages for better debugging.
Summary of changes:
- Removed the redundant check for `image_url` in Ollama format.
- Clarified and enhanced error messages in the conversion logic.
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8232/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8232/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8604
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8604/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8604/comments
|
https://api.github.com/repos/ollama/ollama/issues/8604/events
|
https://github.com/ollama/ollama/pull/8604
| 2,812,382,560
|
PR_kwDOJ0Z1Ps6JDQsP
| 8,604
|
readme: add AI Toolkit for VSCode to community integrations
|
{
"login": "XiaofuHuang",
"id": 49138419,
"node_id": "MDQ6VXNlcjQ5MTM4NDE5",
"avatar_url": "https://avatars.githubusercontent.com/u/49138419?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/XiaofuHuang",
"html_url": "https://github.com/XiaofuHuang",
"followers_url": "https://api.github.com/users/XiaofuHuang/followers",
"following_url": "https://api.github.com/users/XiaofuHuang/following{/other_user}",
"gists_url": "https://api.github.com/users/XiaofuHuang/gists{/gist_id}",
"starred_url": "https://api.github.com/users/XiaofuHuang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/XiaofuHuang/subscriptions",
"organizations_url": "https://api.github.com/users/XiaofuHuang/orgs",
"repos_url": "https://api.github.com/users/XiaofuHuang/repos",
"events_url": "https://api.github.com/users/XiaofuHuang/events{/privacy}",
"received_events_url": "https://api.github.com/users/XiaofuHuang/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2025-01-27T08:34:26
| 2025-01-30T13:12:11
| 2025-01-27T08:36:23
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8604",
"html_url": "https://github.com/ollama/ollama/pull/8604",
"diff_url": "https://github.com/ollama/ollama/pull/8604.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8604.patch",
"merged_at": "2025-01-27T08:36:23"
}
|
Add [AI Toolkit for VSCode](https://marketplace.visualstudio.com/items?itemName=ms-windows-ai-studio.windows-ai-studio) to community integrations.
AI Toolkit for VSCode ollama support: https://aka.ms/ai-tooklit/ollama-docs
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8604/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8604/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2949
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2949/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2949/comments
|
https://api.github.com/repos/ollama/ollama/issues/2949/events
|
https://github.com/ollama/ollama/pull/2949
| 2,171,044,978
|
PR_kwDOJ0Z1Ps5o0Jnk
| 2,949
|
add api.NewClient
|
{
"login": "jackielii",
"id": 360983,
"node_id": "MDQ6VXNlcjM2MDk4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jackielii",
"html_url": "https://github.com/jackielii",
"followers_url": "https://api.github.com/users/jackielii/followers",
"following_url": "https://api.github.com/users/jackielii/following{/other_user}",
"gists_url": "https://api.github.com/users/jackielii/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jackielii/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jackielii/subscriptions",
"organizations_url": "https://api.github.com/users/jackielii/orgs",
"repos_url": "https://api.github.com/users/jackielii/repos",
"events_url": "https://api.github.com/users/jackielii/events{/privacy}",
"received_events_url": "https://api.github.com/users/jackielii/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-03-06T09:31:58
| 2024-05-04T08:35:08
| 2024-05-04T08:35:07
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2949",
"html_url": "https://github.com/ollama/ollama/pull/2949",
"diff_url": "https://github.com/ollama/ollama/pull/2949.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2949.patch",
"merged_at": null
}
|
fix #2948
|
{
"login": "jackielii",
"id": 360983,
"node_id": "MDQ6VXNlcjM2MDk4Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/360983?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jackielii",
"html_url": "https://github.com/jackielii",
"followers_url": "https://api.github.com/users/jackielii/followers",
"following_url": "https://api.github.com/users/jackielii/following{/other_user}",
"gists_url": "https://api.github.com/users/jackielii/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jackielii/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jackielii/subscriptions",
"organizations_url": "https://api.github.com/users/jackielii/orgs",
"repos_url": "https://api.github.com/users/jackielii/repos",
"events_url": "https://api.github.com/users/jackielii/events{/privacy}",
"received_events_url": "https://api.github.com/users/jackielii/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2949/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2949/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1374
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1374/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1374/comments
|
https://api.github.com/repos/ollama/ollama/issues/1374/events
|
https://github.com/ollama/ollama/issues/1374
| 2,024,279,782
|
I_kwDOJ0Z1Ps54qA7m
| 1,374
|
Out of memory error on model that previously worked fine after update to version 0.1.13
|
{
"login": "madsamjp",
"id": 49611363,
"node_id": "MDQ6VXNlcjQ5NjExMzYz",
"avatar_url": "https://avatars.githubusercontent.com/u/49611363?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/madsamjp",
"html_url": "https://github.com/madsamjp",
"followers_url": "https://api.github.com/users/madsamjp/followers",
"following_url": "https://api.github.com/users/madsamjp/following{/other_user}",
"gists_url": "https://api.github.com/users/madsamjp/gists{/gist_id}",
"starred_url": "https://api.github.com/users/madsamjp/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/madsamjp/subscriptions",
"organizations_url": "https://api.github.com/users/madsamjp/orgs",
"repos_url": "https://api.github.com/users/madsamjp/repos",
"events_url": "https://api.github.com/users/madsamjp/events{/privacy}",
"received_events_url": "https://api.github.com/users/madsamjp/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 27
| 2023-12-04T16:40:23
| 2024-01-08T21:42:02
| 2024-01-08T21:42:02
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I configured a model to run entirely in VRAM using the following Modelfile:
```
FROM deepseek-coder:33b-instruct-q5_K_S
PARAMETER num_gpu 65
PARAMETER num_ctx 2048
```
I had no issues with running this, it would use about 22GB of my 4090's 24GB VRAM without issue. It would generate responses very quickly which was very helpful for getting quick answers to short coding queries.
However, yesterday I updated Ollama (to 0.1.13), and now I cannot run the same model. I get an out of memory error, despite the model not needing more than 22.5GB (according to the logs below).
I run Ollama on a headless linux server, so there are no other applications using the GPU.
Was there an update that changes how much VRAM Ollama allocates to make it need more than before? Is there a way to configure Ollama so that it behaves the same way as before?
EDIT: Reverting back to ollama version 0.1.11 resolves the issue for now.
Error:
```
Dec 04 16:28:20 osm-server ollama[528776]: llm_load_tensors: offloaded 65/65 layers to GPU
Dec 04 16:28:20 osm-server ollama[528776]: llm_load_tensors: VRAM used: 21741.89 MiB
Dec 04 16:28:23 osm-server ollama[528776]: ....................................................................................................
Dec 04 16:28:23 osm-server ollama[528776]: llama_new_context_with_model: n_ctx = 2048
Dec 04 16:28:23 osm-server ollama[528776]: llama_new_context_with_model: freq_base = 100000.0
Dec 04 16:28:23 osm-server ollama[528776]: llama_new_context_with_model: freq_scale = 0.25
Dec 04 16:28:24 osm-server ollama[528776]: llama_kv_cache_init: offloading v cache to GPU
Dec 04 16:28:24 osm-server ollama[528776]: llama_kv_cache_init: offloading k cache to GPU
Dec 04 16:28:24 osm-server ollama[528776]: llama_kv_cache_init: VRAM kv self = 496.00 MiB
Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: kv self size = 496.00 MiB
Dec 04 16:28:24 osm-server ollama[528776]: llama_build_graph: non-view tensors processed: 1430/1430
Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: compute buffer total size = 273.07 MiB
Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: VRAM scratch buffer: 270.00 MiB
Dec 04 16:28:24 osm-server ollama[528776]: llama_new_context_with_model: total VRAM used: 22507.89 MiB (model: 21741.89 MiB, context: 766.00 MiB)
Dec 04 16:28:24 osm-server ollama[600735]: {"timestamp":1701707304,"level":"INFO","function":"main","line":2917,"message":"HTTP server listening","hostname":"127.0.0.1","port":57264}
Dec 04 16:28:24 osm-server ollama[600735]: {"timestamp":1701707304,"level":"INFO","function":"log_server_request","line":2478,"message":"request","remote_addr":"127.0.0.1","remote_port":46990,"status":200,"method":"HEAD","path":"/","params":{}}
Dec 04 16:28:24 osm-server ollama[528776]: 2023/12/04 16:28:24 llama.go:493: llama runner started in 4.401485 seconds
Dec 04 16:28:24 osm-server ollama[528776]: CUDA error 2 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:5884: out of memory
Dec 04 16:28:24 osm-server ollama[528776]: current device: 0
Dec 04 16:28:25 osm-server ollama[528776]: 2023/12/04 16:28:25 llama.go:436: 2 at /go/src/github.com/jmorganca/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:5884: out of memory
Dec 04 16:28:25 osm-server ollama[528776]: current device: 0
Dec 04 16:28:25 osm-server ollama[528776]: 2023/12/04 16:28:25 llama.go:510: llama runner stopped successfully
Dec 04 16:28:25 osm-server ollama[528776]: [GIN] 2023/12/04 - 16:28:25 | 200 | 6.468638351s | 127.0.0.1 | POST "/api/generate"
```
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1374/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1374/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/938
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/938/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/938/comments
|
https://api.github.com/repos/ollama/ollama/issues/938/events
|
https://github.com/ollama/ollama/issues/938
| 1,966,217,268
|
I_kwDOJ0Z1Ps51Mhg0
| 938
|
CLI windows size detection not accurate on WSL
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 0
| 2023-10-27T21:33:22
| 2024-02-20T00:56:13
| 2024-02-20T00:56:13
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
When backspacing characters from a long input on wsl a "float extra character" may not be deleted.
Reproduce by going to a newline and backspacing from the last character.
<img width="717" alt="Screenshot 2023-10-27 at 17 30 48" src="https://github.com/jmorganca/ollama/assets/5853428/9e18a84b-3071-4b67-aaba-43f2e2d4c1f4">
<img width="716" alt="Screenshot 2023-10-27 at 17 31 12" src="https://github.com/jmorganca/ollama/assets/5853428/029e7b25-d49a-435e-b320-773a22d81369">
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/938/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/938/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3458
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3458/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3458/comments
|
https://api.github.com/repos/ollama/ollama/issues/3458/events
|
https://github.com/ollama/ollama/pull/3458
| 2,220,673,215
|
PR_kwDOJ0Z1Ps5rcjJg
| 3,458
|
fix model needLoad always be true
|
{
"login": "mofanke",
"id": 54242816,
"node_id": "MDQ6VXNlcjU0MjQyODE2",
"avatar_url": "https://avatars.githubusercontent.com/u/54242816?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mofanke",
"html_url": "https://github.com/mofanke",
"followers_url": "https://api.github.com/users/mofanke/followers",
"following_url": "https://api.github.com/users/mofanke/following{/other_user}",
"gists_url": "https://api.github.com/users/mofanke/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mofanke/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mofanke/subscriptions",
"organizations_url": "https://api.github.com/users/mofanke/orgs",
"repos_url": "https://api.github.com/users/mofanke/repos",
"events_url": "https://api.github.com/users/mofanke/events{/privacy}",
"received_events_url": "https://api.github.com/users/mofanke/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-04-02T14:26:53
| 2024-04-03T06:05:55
| 2024-04-03T06:05:54
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3458",
"html_url": "https://github.com/ollama/ollama/pull/3458",
"diff_url": "https://github.com/ollama/ollama/pull/3458.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3458.patch",
"merged_at": null
}
|

loaded.Options.Runner.NumGPU was change to 33 or other value by
``
func New(model string, adapters, projectors []string, opts *api.Options)
``
so `!reflect.DeepEqual(loaded.Options.Runner, opts.Runner)` will always be true, because` opts.Runner.NumGPU` which default to -1
|
{
"login": "mofanke",
"id": 54242816,
"node_id": "MDQ6VXNlcjU0MjQyODE2",
"avatar_url": "https://avatars.githubusercontent.com/u/54242816?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mofanke",
"html_url": "https://github.com/mofanke",
"followers_url": "https://api.github.com/users/mofanke/followers",
"following_url": "https://api.github.com/users/mofanke/following{/other_user}",
"gists_url": "https://api.github.com/users/mofanke/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mofanke/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mofanke/subscriptions",
"organizations_url": "https://api.github.com/users/mofanke/orgs",
"repos_url": "https://api.github.com/users/mofanke/repos",
"events_url": "https://api.github.com/users/mofanke/events{/privacy}",
"received_events_url": "https://api.github.com/users/mofanke/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3458/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3458/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/729
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/729/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/729/comments
|
https://api.github.com/repos/ollama/ollama/issues/729/events
|
https://github.com/ollama/ollama/issues/729
| 1,931,317,404
|
I_kwDOJ0Z1Ps5zHZCc
| 729
|
Unable to pull models behind the proxy
|
{
"login": "ilyanoskov",
"id": 16443383,
"node_id": "MDQ6VXNlcjE2NDQzMzgz",
"avatar_url": "https://avatars.githubusercontent.com/u/16443383?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ilyanoskov",
"html_url": "https://github.com/ilyanoskov",
"followers_url": "https://api.github.com/users/ilyanoskov/followers",
"following_url": "https://api.github.com/users/ilyanoskov/following{/other_user}",
"gists_url": "https://api.github.com/users/ilyanoskov/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ilyanoskov/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ilyanoskov/subscriptions",
"organizations_url": "https://api.github.com/users/ilyanoskov/orgs",
"repos_url": "https://api.github.com/users/ilyanoskov/repos",
"events_url": "https://api.github.com/users/ilyanoskov/events{/privacy}",
"received_events_url": "https://api.github.com/users/ilyanoskov/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 13
| 2023-10-07T11:38:39
| 2024-10-01T11:45:58
| 2023-10-10T16:59:07
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Dear Maintainers,
Thank you very much for creating this project!
I need to set up ollama on Linux behind a proxy, and when pulling I get an error:
```download.go:166: couldn't download blob: Get "https:///...../ollama/docker/registry/v2/blobs/...": tls: first record does not look like a TLS handshake```
I have tried these methods and they also did not work for me:
- https://github.com/jmorganca/ollama/issues/703#issuecomment-1747857562
- https://github.com/jmorganca/ollama/issues/676#issuecomment-1744722380
- https://github.com/jmorganca/ollama/issues/697
Could you please add a way to configure a proxy for Ollama? That will enable a lot of users that must use a proxy.
Thank you very much in advance.
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/729/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/729/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4922
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4922/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4922/comments
|
https://api.github.com/repos/ollama/ollama/issues/4922/events
|
https://github.com/ollama/ollama/pull/4922
| 2,341,320,564
|
PR_kwDOJ0Z1Ps5x2EeC
| 4,922
|
cmd: initial functionality for `ollama show <model>`
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-06-08T00:44:51
| 2024-06-15T03:03:48
| 2024-06-14T23:18:33
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | true
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4922",
"html_url": "https://github.com/ollama/ollama/pull/4922",
"diff_url": "https://github.com/ollama/ollama/pull/4922.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4922.patch",
"merged_at": null
}
| null |
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4922/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4922/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4631
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4631/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4631/comments
|
https://api.github.com/repos/ollama/ollama/issues/4631/events
|
https://github.com/ollama/ollama/issues/4631
| 2,316,815,049
|
I_kwDOJ0Z1Ps6KF8rJ
| 4,631
|
Repeating answers in an instance.
|
{
"login": "Protostarship",
"id": 125519972,
"node_id": "U_kgDOB3tIZA",
"avatar_url": "https://avatars.githubusercontent.com/u/125519972?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Protostarship",
"html_url": "https://github.com/Protostarship",
"followers_url": "https://api.github.com/users/Protostarship/followers",
"following_url": "https://api.github.com/users/Protostarship/following{/other_user}",
"gists_url": "https://api.github.com/users/Protostarship/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Protostarship/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Protostarship/subscriptions",
"organizations_url": "https://api.github.com/users/Protostarship/orgs",
"repos_url": "https://api.github.com/users/Protostarship/repos",
"events_url": "https://api.github.com/users/Protostarship/events{/privacy}",
"received_events_url": "https://api.github.com/users/Protostarship/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 1
| 2024-05-25T07:58:33
| 2024-05-29T22:52:59
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hi, i'm using llama2 models, and when i asked the Ai to explain something, it does responds and explain it, but when it reached the end, it instead print out certain line,
```
[INST: <</INST]
```
then repeats another answer regarding the current questions.
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
ollama 0.1.38
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4631/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4631/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/6263
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6263/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6263/comments
|
https://api.github.com/repos/ollama/ollama/issues/6263/events
|
https://github.com/ollama/ollama/issues/6263
| 2,456,652,905
|
I_kwDOJ0Z1Ps6SbYxp
| 6,263
|
Pull Command Parsing Not Working
|
{
"login": "chadwickhar08",
"id": 23239905,
"node_id": "MDQ6VXNlcjIzMjM5OTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/23239905?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chadwickhar08",
"html_url": "https://github.com/chadwickhar08",
"followers_url": "https://api.github.com/users/chadwickhar08/followers",
"following_url": "https://api.github.com/users/chadwickhar08/following{/other_user}",
"gists_url": "https://api.github.com/users/chadwickhar08/gists{/gist_id}",
"starred_url": "https://api.github.com/users/chadwickhar08/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/chadwickhar08/subscriptions",
"organizations_url": "https://api.github.com/users/chadwickhar08/orgs",
"repos_url": "https://api.github.com/users/chadwickhar08/repos",
"events_url": "https://api.github.com/users/chadwickhar08/events{/privacy}",
"received_events_url": "https://api.github.com/users/chadwickhar08/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 9
| 2024-08-08T21:27:30
| 2024-08-10T00:18:27
| 2024-08-09T19:30:26
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
When running ollama on Windows, attempt to run 'ollama pull llama3.1' results in 'ollama pull llama3.1
pulling manifest
Error: Incorrect function.'
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
ollama --version ollama version is 0.3.4
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6263/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6263/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6412
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6412/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6412/comments
|
https://api.github.com/repos/ollama/ollama/issues/6412/events
|
https://github.com/ollama/ollama/issues/6412
| 2,472,517,560
|
I_kwDOJ0Z1Ps6TX5-4
| 6,412
|
for glm4-9b
|
{
"login": "lcck200121",
"id": 126126214,
"node_id": "U_kgDOB4SIhg",
"avatar_url": "https://avatars.githubusercontent.com/u/126126214?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lcck200121",
"html_url": "https://github.com/lcck200121",
"followers_url": "https://api.github.com/users/lcck200121/followers",
"following_url": "https://api.github.com/users/lcck200121/following{/other_user}",
"gists_url": "https://api.github.com/users/lcck200121/gists{/gist_id}",
"starred_url": "https://api.github.com/users/lcck200121/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lcck200121/subscriptions",
"organizations_url": "https://api.github.com/users/lcck200121/orgs",
"repos_url": "https://api.github.com/users/lcck200121/repos",
"events_url": "https://api.github.com/users/lcck200121/events{/privacy}",
"received_events_url": "https://api.github.com/users/lcck200121/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-08-19T05:54:49
| 2024-08-19T05:54:49
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Do you have any plans to support the tool-calling function of glm4-9b?
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6412/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6412/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/4194
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4194/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4194/comments
|
https://api.github.com/repos/ollama/ollama/issues/4194/events
|
https://github.com/ollama/ollama/pull/4194
| 2,280,062,512
|
PR_kwDOJ0Z1Ps5umAi_
| 4,194
|
docs: delete log `HEAD`
|
{
"login": "HydenLiu",
"id": 52030677,
"node_id": "MDQ6VXNlcjUyMDMwNjc3",
"avatar_url": "https://avatars.githubusercontent.com/u/52030677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/HydenLiu",
"html_url": "https://github.com/HydenLiu",
"followers_url": "https://api.github.com/users/HydenLiu/followers",
"following_url": "https://api.github.com/users/HydenLiu/following{/other_user}",
"gists_url": "https://api.github.com/users/HydenLiu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/HydenLiu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/HydenLiu/subscriptions",
"organizations_url": "https://api.github.com/users/HydenLiu/orgs",
"repos_url": "https://api.github.com/users/HydenLiu/repos",
"events_url": "https://api.github.com/users/HydenLiu/events{/privacy}",
"received_events_url": "https://api.github.com/users/HydenLiu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-05-06T05:28:47
| 2024-05-06T17:32:31
| 2024-05-06T17:32:31
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4194",
"html_url": "https://github.com/ollama/ollama/pull/4194",
"diff_url": "https://github.com/ollama/ollama/pull/4194.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4194.patch",
"merged_at": "2024-05-06T17:32:31"
}
| null |
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4194/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4194/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1804
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1804/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1804/comments
|
https://api.github.com/repos/ollama/ollama/issues/1804/events
|
https://github.com/ollama/ollama/issues/1804
| 2,067,249,992
|
I_kwDOJ0Z1Ps57N7tI
| 1,804
|
Code view on codellama vs phi and dolphin-phi
|
{
"login": "oliverbob",
"id": 23272429,
"node_id": "MDQ6VXNlcjIzMjcyNDI5",
"avatar_url": "https://avatars.githubusercontent.com/u/23272429?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/oliverbob",
"html_url": "https://github.com/oliverbob",
"followers_url": "https://api.github.com/users/oliverbob/followers",
"following_url": "https://api.github.com/users/oliverbob/following{/other_user}",
"gists_url": "https://api.github.com/users/oliverbob/gists{/gist_id}",
"starred_url": "https://api.github.com/users/oliverbob/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/oliverbob/subscriptions",
"organizations_url": "https://api.github.com/users/oliverbob/orgs",
"repos_url": "https://api.github.com/users/oliverbob/repos",
"events_url": "https://api.github.com/users/oliverbob/events{/privacy}",
"received_events_url": "https://api.github.com/users/oliverbob/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-01-05T12:19:46
| 2024-03-13T23:53:54
| 2024-03-13T23:53:54
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Dolphin phi and (probably phi code indent):


Phi, indents but has no code view:

Can someone probably do something to improve these models as they are the only models that run very fast on smaller GPUs.
Or perhaps, maybe someone would train phi-code:instruct.
Thanks.
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1804/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1804/timeline
| null |
completed
| false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.