url
stringlengths 51
54
| repository_url
stringclasses 1
value | labels_url
stringlengths 65
68
| comments_url
stringlengths 60
63
| events_url
stringlengths 58
61
| html_url
stringlengths 39
44
| id
int64 1.78B
2.82B
| node_id
stringlengths 18
19
| number
int64 1
8.69k
| title
stringlengths 1
382
| user
dict | labels
listlengths 0
5
| state
stringclasses 2
values | locked
bool 1
class | assignee
dict | assignees
listlengths 0
2
| milestone
null | comments
int64 0
323
| created_at
timestamp[s] | updated_at
timestamp[s] | closed_at
timestamp[s] | author_association
stringclasses 4
values | sub_issues_summary
dict | active_lock_reason
null | draft
bool 2
classes | pull_request
dict | body
stringlengths 2
118k
⌀ | closed_by
dict | reactions
dict | timeline_url
stringlengths 60
63
| performed_via_github_app
null | state_reason
stringclasses 4
values | is_pull_request
bool 2
classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
https://api.github.com/repos/ollama/ollama/issues/1718
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1718/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1718/comments
|
https://api.github.com/repos/ollama/ollama/issues/1718/events
|
https://github.com/ollama/ollama/issues/1718
| 2,056,056,305
|
I_kwDOJ0Z1Ps56jO3x
| 1,718
|
incomplete json in api responses
|
{
"login": "ralyodio",
"id": 27381,
"node_id": "MDQ6VXNlcjI3Mzgx",
"avatar_url": "https://avatars.githubusercontent.com/u/27381?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ralyodio",
"html_url": "https://github.com/ralyodio",
"followers_url": "https://api.github.com/users/ralyodio/followers",
"following_url": "https://api.github.com/users/ralyodio/following{/other_user}",
"gists_url": "https://api.github.com/users/ralyodio/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ralyodio/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ralyodio/subscriptions",
"organizations_url": "https://api.github.com/users/ralyodio/orgs",
"repos_url": "https://api.github.com/users/ralyodio/repos",
"events_url": "https://api.github.com/users/ralyodio/events{/privacy}",
"received_events_url": "https://api.github.com/users/ralyodio/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2023-12-26T05:51:29
| 2023-12-26T17:22:36
| 2023-12-26T17:22:36
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I tried both /api/chat and /api/generate endpoints which seem to produce the same results. however I'm getting invalid json on every response.
|
{
"login": "ralyodio",
"id": 27381,
"node_id": "MDQ6VXNlcjI3Mzgx",
"avatar_url": "https://avatars.githubusercontent.com/u/27381?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ralyodio",
"html_url": "https://github.com/ralyodio",
"followers_url": "https://api.github.com/users/ralyodio/followers",
"following_url": "https://api.github.com/users/ralyodio/following{/other_user}",
"gists_url": "https://api.github.com/users/ralyodio/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ralyodio/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ralyodio/subscriptions",
"organizations_url": "https://api.github.com/users/ralyodio/orgs",
"repos_url": "https://api.github.com/users/ralyodio/repos",
"events_url": "https://api.github.com/users/ralyodio/events{/privacy}",
"received_events_url": "https://api.github.com/users/ralyodio/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1718/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1718/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7626
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7626/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7626/comments
|
https://api.github.com/repos/ollama/ollama/issues/7626/events
|
https://github.com/ollama/ollama/issues/7626
| 2,651,476,994
|
I_kwDOJ0Z1Ps6eClQC
| 7,626
|
Role field should not be repeated in streamed response chunks
|
{
"login": "jackmpcollins",
"id": 6640905,
"node_id": "MDQ6VXNlcjY2NDA5MDU=",
"avatar_url": "https://avatars.githubusercontent.com/u/6640905?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jackmpcollins",
"html_url": "https://github.com/jackmpcollins",
"followers_url": "https://api.github.com/users/jackmpcollins/followers",
"following_url": "https://api.github.com/users/jackmpcollins/following{/other_user}",
"gists_url": "https://api.github.com/users/jackmpcollins/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jackmpcollins/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jackmpcollins/subscriptions",
"organizations_url": "https://api.github.com/users/jackmpcollins/orgs",
"repos_url": "https://api.github.com/users/jackmpcollins/repos",
"events_url": "https://api.github.com/users/jackmpcollins/events{/privacy}",
"received_events_url": "https://api.github.com/users/jackmpcollins/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 7706482389,
"node_id": "LA_kwDOJ0Z1Ps8AAAABy1eW1Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/api",
"name": "api",
"color": "bfdadc",
"default": false,
"description": ""
}
] |
open
| false
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1
| 2024-11-12T08:46:18
| 2024-11-18T07:52:26
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
The streamed chat-completion response from ollama's openai-compatible API repeats `"role": "assistant"` in all returned chunks. This is different to OpenAI's API which just has this in the first chunk. This breaks compatibility with the `client.beta.chat.completions.stream` helper from the openai package. See also this issue https://github.com/pydantic/logfire/pull/545#discussion_r1837660027. Ollama should omit the "role" field or return `"role": None` for all chunks after the first one.
---
OpenAI chunks: "role" only in first chunk
```python
from openai import OpenAI
client = Client()
response = client.chat.completions.create(
model="gpt-4",
messages=[{"role": "user", "content": "Just say: The answer is secret."}],
stream=True,
)
for chunk in response:
print(chunk.model_dump_json(exclude_none=True))
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{"content":"","role":"assistant"},"index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{"content":"The"},"index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{"content":" answer"},"index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{"content":" is"},"index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{"content":" secret"},"index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{"content":"."},"index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
{"id":"chatcmpl-ASgaImINIA8gwsca92CCgES2VldF8","choices":[{"delta":{},"finish_reason":"stop","index":0}],"created":1731400242,"model":"gpt-4-0613","object":"chat.completion.chunk"}
```
Ollama chunks: "role" provided in every chunk
```python
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:11434/v1",
api_key="ollama",
)
response = client.chat.completions.create(
model="llama3.1",
messages=[{"role": "user", "content": "Just say: The answer is secret."}],
stream=True,
# stream_options={"include_usage": True},
# max_tokens=1,
)
for chunk in response:
print(chunk.model_dump_json(exclude_none=True))
{"id":"chatcmpl-230","choices":[{"delta":{"content":"The","role":"assistant"},"index":0}],"created":1731400290,"model":"llama3.1","object":"chat.completion.chunk","system_fingerprint":"fp_ollama"}
{"id":"chatcmpl-230","choices":[{"delta":{"content":" answer","role":"assistant"},"index":0}],"created":1731400290,"model":"llama3.1","object":"chat.completion.chunk","system_fingerprint":"fp_ollama"}
{"id":"chatcmpl-230","choices":[{"delta":{"content":" is","role":"assistant"},"index":0}],"created":1731400290,"model":"llama3.1","object":"chat.completion.chunk","system_fingerprint":"fp_ollama"}
{"id":"chatcmpl-230","choices":[{"delta":{"content":" secret","role":"assistant"},"index":0}],"created":1731400290,"model":"llama3.1","object":"chat.completion.chunk","system_fingerprint":"fp_ollama"}
{"id":"chatcmpl-230","choices":[{"delta":{"content":".","role":"assistant"},"index":0}],"created":1731400290,"model":"llama3.1","object":"chat.completion.chunk","system_fingerprint":"fp_ollama"}
{"id":"chatcmpl-230","choices":[{"delta":{"content":"","role":"assistant"},"finish_reason":"stop","index":0}],"created":1731400290,"model":"llama3.1","object":"chat.completion.chunk","system_fingerprint":"fp_ollama"}
```
Using `client.beta.chat.completions.stream` with ollama results in `"role": "assistantassistant...`.
openai docs: https://github.com/openai/openai-python/blob/646a579cdb305a9d3fba6c5f9a96011c5e2c2882/helpers.md#chat-completions-api
```python
from openai import OpenAI
client = OpenAI(
base_url="http://localhost:11434/v1",
api_key="ollama",
)
with client.beta.chat.completions.stream(
model="llama3.1",
messages=[{"role": "user", "content": "Just say: The answer is secret."}],
) as stream:
for event in stream:
pass
print(stream.get_final_completion().model_dump_json(indent=2))
{
"id": "chatcmpl-653",
"choices": [
{
"finish_reason": "stop",
"index": 0,
"logprobs": null,
"message": {
"content": "The answer is secret.",
"refusal": null,
"role": "assistantassistantassistantassistantassistantassistant",
"audio": null,
"function_call": null,
"tool_calls": [],
"parsed": null
}
}
],
"created": 1731400312,
"model": "llama3.1",
"object": "chat.completion",
"service_tier": null,
"system_fingerprint": "fp_ollama",
"usage": null
}
```
### OS
macOS
### GPU
_No response_
### CPU
_No response_
### Ollama version
0.4.1
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7626/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7626/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/3868
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3868/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3868/comments
|
https://api.github.com/repos/ollama/ollama/issues/3868/events
|
https://github.com/ollama/ollama/issues/3868
| 2,260,448,866
|
I_kwDOJ0Z1Ps6Gu7Zi
| 3,868
|
Hope to get it out on the shelves llama3-Chinese
|
{
"login": "enryteam",
"id": 20081090,
"node_id": "MDQ6VXNlcjIwMDgxMDkw",
"avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/enryteam",
"html_url": "https://github.com/enryteam",
"followers_url": "https://api.github.com/users/enryteam/followers",
"following_url": "https://api.github.com/users/enryteam/following{/other_user}",
"gists_url": "https://api.github.com/users/enryteam/gists{/gist_id}",
"starred_url": "https://api.github.com/users/enryteam/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/enryteam/subscriptions",
"organizations_url": "https://api.github.com/users/enryteam/orgs",
"repos_url": "https://api.github.com/users/enryteam/repos",
"events_url": "https://api.github.com/users/enryteam/events{/privacy}",
"received_events_url": "https://api.github.com/users/enryteam/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 0
| 2024-04-24T06:23:11
| 2024-07-20T14:29:51
| 2024-07-20T14:29:51
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
比如 https://github.com/UnicomAI/Unichat-llama3-Chinese
|
{
"login": "enryteam",
"id": 20081090,
"node_id": "MDQ6VXNlcjIwMDgxMDkw",
"avatar_url": "https://avatars.githubusercontent.com/u/20081090?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/enryteam",
"html_url": "https://github.com/enryteam",
"followers_url": "https://api.github.com/users/enryteam/followers",
"following_url": "https://api.github.com/users/enryteam/following{/other_user}",
"gists_url": "https://api.github.com/users/enryteam/gists{/gist_id}",
"starred_url": "https://api.github.com/users/enryteam/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/enryteam/subscriptions",
"organizations_url": "https://api.github.com/users/enryteam/orgs",
"repos_url": "https://api.github.com/users/enryteam/repos",
"events_url": "https://api.github.com/users/enryteam/events{/privacy}",
"received_events_url": "https://api.github.com/users/enryteam/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3868/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3868/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4333
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4333/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4333/comments
|
https://api.github.com/repos/ollama/ollama/issues/4333/events
|
https://github.com/ollama/ollama/issues/4333
| 2,290,617,478
|
I_kwDOJ0Z1Ps6IiAyG
| 4,333
|
`segmentation fault` when running `codellama:34b` on A100
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6430601766,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf0syJg",
"url": "https://api.github.com/repos/ollama/ollama/labels/nvidia",
"name": "nvidia",
"color": "8CDB00",
"default": false,
"description": "Issues relating to Nvidia GPUs and CUDA"
},
{
"id": 6677745918,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgZQ_g",
"url": "https://api.github.com/repos/ollama/ollama/labels/gpu",
"name": "gpu",
"color": "76C49E",
"default": false,
"description": ""
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 6
| 2024-05-11T02:48:28
| 2024-07-22T18:05:25
| 2024-07-22T18:05:25
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
CLI:
```
$ ollama run codellama:34b
Error: llama runner process has terminated: signal: segmentation fault
```
Logs:
```
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.033Z level=INFO source=memory.go:127 msg="offload to gpu" layers.real=-1 layers.estimate=49 memory.available="39.0 GiB" memory.required.full="19.1 GiB" memory.required.partial="19.1 GiB" memory.required.kv="384.0 MiB" memory.weights.total="18.0 GiB" memory.weights.repeating="17.8 GiB" memory.weights.nonrepeating="205.1 MiB" memory.graph.full="324.0 MiB" memory.graph.partial="348.0 MiB"
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.034Z level=INFO source=memory.go:127 msg="offload to gpu" layers.real=-1 layers.estimate=49 memory.available="39.0 GiB" memory.required.full="19.1 GiB" memory.required.partial="19.1 GiB" memory.required.kv="384.0 MiB" memory.weights.total="18.0 GiB" memory.weights.repeating="17.8 GiB" memory.weights.nonrepeating="205.1 MiB" memory.graph.full="324.0 MiB" memory.graph.partial="348.0 MiB"
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.035Z level=INFO source=server.go:308 msg="starting llama server" cmd="/tmp/ollama944909272/runners/cuda_v11/ollama_llama_server --model /usr/share/ollama/.ollama/models/blobs/sha256-f36b668ebcd329357fac22db35f6414a1c9309307f33d08fe217bbf84b0496cc --ctx-size 2048 --batch-size 512 --embedding --log-disable --n-gpu-layers 49 --parallel 1 --port 36091"
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.035Z level=INFO source=sched.go:333 msg="loaded runners" count=1
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.035Z level=INFO source=server.go:478 msg="waiting for llama runner to start responding"
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.035Z level=INFO source=server.go:514 msg="waiting for server to become available" status="llm server error"
May 11 02:47:28 gpu ollama[28220]: INFO [main] build info | build=1 commit="952d03d" tid="140151386750976" timestamp=1715395648
May 11 02:47:28 gpu ollama[28220]: INFO [main] system info | n_threads=6 n_threads_batch=-1 system_info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | " tid="140151386750976" timestamp=1715395648 total_threads=12
May 11 02:47:28 gpu ollama[28220]: INFO [main] HTTP server listening | hostname="127.0.0.1" n_threads_http="11" port="36091" tid="140151386750976" timestamp=1715395648
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: loaded meta data with 20 key-value pairs and 435 tensors from /usr/share/ollama/.ollama/models/blobs/sha256-f36b668ebcd329357fac22db35f6414a1c9309307f33d08fe217bbf84b0496cc (version GGUF V2)
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 0: general.architecture str = llama
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 1: general.name str = codellama
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 2: llama.context_length u32 = 16384
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 3: llama.embedding_length u32 = 8192
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 4: llama.block_count u32 = 48
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 5: llama.feed_forward_length u32 = 22016
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 6: llama.rope.dimension_count u32 = 128
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 7: llama.attention.head_count u32 = 64
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 8: llama.attention.head_count_kv u32 = 8
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 9: llama.attention.layer_norm_rms_epsilon f32 = 0.000010
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 10: llama.rope.freq_base f32 = 1000000.000000
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 11: general.file_type u32 = 2
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 12: tokenizer.ggml.model str = llama
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 13: tokenizer.ggml.tokens arr[str,32000] = ["<unk>", "<s>", "</s>", "<0x00>", "<...
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 14: tokenizer.ggml.scores arr[f32,32000] = [0.000000, 0.000000, 0.000000, 0.0000...
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 15: tokenizer.ggml.token_type arr[i32,32000] = [2, 3, 3, 6, 6, 6, 6, 6, 6, 6, 6, 6, ...
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 16: tokenizer.ggml.bos_token_id u32 = 1
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 17: tokenizer.ggml.eos_token_id u32 = 2
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 18: tokenizer.ggml.unknown_token_id u32 = 0
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - kv 19: general.quantization_version u32 = 2
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - type f32: 97 tensors
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - type q4_0: 337 tensors
May 11 02:47:28 gpu ollama[27286]: llama_model_loader: - type q6_K: 1 tensors
May 11 02:47:28 gpu ollama[27286]: llm_load_vocab: special tokens definition check successful ( 259/32000 ).
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: format = GGUF V2
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: arch = llama
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: vocab type = SPM
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_vocab = 32000
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_merges = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_ctx_train = 16384
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_embd = 8192
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_head = 64
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_head_kv = 8
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_layer = 48
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_rot = 128
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_embd_head_k = 128
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_embd_head_v = 128
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_gqa = 8
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_embd_k_gqa = 1024
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_embd_v_gqa = 1024
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: f_norm_eps = 0.0e+00
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: f_norm_rms_eps = 1.0e-05
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: f_clamp_kqv = 0.0e+00
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: f_max_alibi_bias = 0.0e+00
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: f_logit_scale = 0.0e+00
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_ff = 22016
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_expert = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_expert_used = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: causal attn = 1
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: pooling type = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: rope type = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: rope scaling = linear
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: freq_base_train = 1000000.0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: freq_scale_train = 1
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: n_yarn_orig_ctx = 16384
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: rope_finetuned = unknown
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: ssm_d_conv = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: ssm_d_inner = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: ssm_d_state = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: ssm_dt_rank = 0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: model type = 34B
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: model ftype = Q4_0
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: model params = 33.74 B
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: model size = 17.74 GiB (4.52 BPW)
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: general.name = codellama
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: BOS token = 1 '<s>'
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: EOS token = 2 '</s>'
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: UNK token = 0 '<unk>'
May 11 02:47:28 gpu ollama[27286]: llm_load_print_meta: LF token = 13 '<0x0A>'
May 11 02:47:28 gpu ollama[27286]: [52B blob data]
May 11 02:47:28 gpu ollama[27286]: time=2024-05-11T02:47:28.286Z level=ERROR source=sched.go:339 msg="error loading llama server" error="llama runner process has terminated: signal: segmentation fault "
May 11 02:47:28 gpu ollama[27286]: [GIN] 2024/05/11 - 02:47:28 | 500 | 1.242539308s | 127.0.0.1 | POST "/api/chat"
May 11 02:47:30 gpu ollama[27286]: time=2024-05-11T02:47:30.881Z level=WARN source=sched.go:507 msg="gpu VRAM usage didn't recover within timeout" seconds=5.22935882
May 11 02:47:31 gpu ollama[27286]: time=2024-05-11T02:47:31.211Z level=WARN source=sched.go:507 msg="gpu VRAM usage didn't recover within timeout" seconds=5.559953212
May 11 02:47:31 gpu ollama[27286]: time=2024-05-11T02:47:31.542Z level=WARN source=sched.go:507 msg="gpu VRAM usage didn't recover within timeout" seconds=5.890430607
May 11 02:47:36 gpu ollama[27286]: time=2024-05-11T02:47:36.102Z level=WARN source=sched.go:507 msg="gpu VRAM usage didn't recover within timeout" seconds=5.221611086
May 11 02:47:36 gpu ollama[27286]: time=2024-05-11T02:47:36.434Z level=WARN source=sched.go:507 msg="gpu VRAM usage didn't recover within timeout" seconds=5.552699521
May 11 02:47:36 gpu ollama[27286]: time=2024-05-11T02:47:36.764Z level=WARN source=sched.go:507 msg="gpu VRAM usage didn't recover within timeout" seconds=5.883013969
```
### OS
Linux
### GPU
NVIDIA A100 40GB
### CPU
Intel
### Ollama version
0.1.35
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4333/reactions",
"total_count": 8,
"+1": 8,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4333/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8443
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8443/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8443/comments
|
https://api.github.com/repos/ollama/ollama/issues/8443/events
|
https://github.com/ollama/ollama/pull/8443
| 2,790,742,559
|
PR_kwDOJ0Z1Ps6H5iUB
| 8,443
|
llama/llama-mmap: fix missing include
|
{
"login": "wgottwalt",
"id": 12194808,
"node_id": "MDQ6VXNlcjEyMTk0ODA4",
"avatar_url": "https://avatars.githubusercontent.com/u/12194808?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wgottwalt",
"html_url": "https://github.com/wgottwalt",
"followers_url": "https://api.github.com/users/wgottwalt/followers",
"following_url": "https://api.github.com/users/wgottwalt/following{/other_user}",
"gists_url": "https://api.github.com/users/wgottwalt/gists{/gist_id}",
"starred_url": "https://api.github.com/users/wgottwalt/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/wgottwalt/subscriptions",
"organizations_url": "https://api.github.com/users/wgottwalt/orgs",
"repos_url": "https://api.github.com/users/wgottwalt/repos",
"events_url": "https://api.github.com/users/wgottwalt/events{/privacy}",
"received_events_url": "https://api.github.com/users/wgottwalt/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 0
| 2025-01-15T20:04:49
| 2025-01-15T20:04:49
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8443",
"html_url": "https://github.com/ollama/ollama/pull/8443",
"diff_url": "https://github.com/ollama/ollama/pull/8443.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8443.patch",
"merged_at": null
}
|
Proper memory and vector headers (like in GCC 15.1) do not provide the uint32_t type, so cstdint is required.
llama-mmap.h:55:20: error: ‘uint32_t’ has not been declared
55 | void write_u32(uint32_t val) const;
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8443/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8443/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7597
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7597/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7597/comments
|
https://api.github.com/repos/ollama/ollama/issues/7597/events
|
https://github.com/ollama/ollama/issues/7597
| 2,647,257,116
|
I_kwDOJ0Z1Ps6dyfAc
| 7,597
|
detect missing GPU runners and don't report incorrect GPU info/logs
|
{
"login": "kaleocheng",
"id": 7939352,
"node_id": "MDQ6VXNlcjc5MzkzNTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/7939352?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kaleocheng",
"html_url": "https://github.com/kaleocheng",
"followers_url": "https://api.github.com/users/kaleocheng/followers",
"following_url": "https://api.github.com/users/kaleocheng/following{/other_user}",
"gists_url": "https://api.github.com/users/kaleocheng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kaleocheng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kaleocheng/subscriptions",
"organizations_url": "https://api.github.com/users/kaleocheng/orgs",
"repos_url": "https://api.github.com/users/kaleocheng/repos",
"events_url": "https://api.github.com/users/kaleocheng/events{/privacy}",
"received_events_url": "https://api.github.com/users/kaleocheng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 20
| 2024-11-10T13:41:47
| 2024-11-17T20:18:32
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
```
$ ollama -v
ollama version is 0.4.1
$ ollama run llama3.2-vision:latest
$ ollama ps
NAME ID SIZE PROCESSOR UNTIL
llama3.2-vision:latest 38107a0cd119 12 GB 100% GPU 2 minutes from now
```
from the logs it also saying ollama offload to cuda:
```
ollama[1773]: [GIN] 2024/11/10 - 21:32:56 | 200 | 22.078108ms | 127.0.0.1 | POST "/api/show"
ollama[1773]: time=2024-11-10T21:32:56.205+08:00 level=WARN source=sched.go:137 msg="mllama doesn't support parallel requests yet"
ollama[1773]: time=2024-11-10T21:32:56.342+08:00 level=INFO source=sched.go:714 msg="new model will fit in available VRAM in single GPU, loading" model=/var/lib/ollama/models/blobs/sha256-11f274007f093fefeec994a5dbbb33d0733a4feb87f7ab66dcd7c1069fef0068 gpu=GPU-957abb1f-e95c-db43-ee81-b345b6e60491 parallel=1 available=16139026432 required="11.3 GiB"
ollama[1773]: time=2024-11-10T21:32:56.440+08:00 level=INFO source=server.go:105 msg="system memory" total="15.4 GiB" free="11.3 GiB" free_swap="12.2 GiB"
ollama[1773]: time=2024-11-10T21:32:56.442+08:00 level=INFO source=memory.go:343 msg="offload to cuda" projector.weights="1.8 GiB" projector.graph="2.8 GiB" layers.requested=-1 layers.model=41 layers.offload=41 layers.split="" memory.available="[15.0 GiB]" memory.gpu_overhead="0 B" memory.required.full="11.3 GiB" memory.required.partial="11.3 GiB" memory.required.kv="656.2 MiB" memory.required.allocations="[11.3 GiB]" memory.weights.total="5.5 GiB" memory.weights.repeating="5.1 GiB" memory.weights.nonrepeating="411.0 MiB" memory.graph.full="258.5 MiB" memory.graph.partial="669.5 MiB"
ollama[1773]: time=2024-11-10T21:32:56.443+08:00 level=INFO source=server.go:383 msg="starting llama server" cmd="/tmp/ollama1704822012/runners/cpu_avx2/ollama_llama_server --model /var/lib/ollama/models/blobs/sha256-11f274007f093fefeec994a5dbbb33d0733a4feb87f7ab66dcd7c1069fef0068 --ctx-size 2048 --batch-size 512 --n-gpu-layers 41 --mmproj /var/lib/ollama/models/blobs/sha256-ece5e659647a20a5c28ab9eea1c12a1ad430bc0f2a27021d00ad103b3bf5206f --threads 6 --no-mmap --parallel 1 --port 40225"
ollama[1773]: time=2024-11-10T21:32:56.443+08:00 level=INFO source=sched.go:449 msg="loaded runners" count=1
ollama[1773]: time=2024-11-10T21:32:56.443+08:00 level=INFO source=server.go:562 msg="waiting for llama runner to start responding"
ollama[1773]: time=2024-11-10T21:32:56.444+08:00 level=INFO source=server.go:596 msg="waiting for server to become available" status="llm server error"
ollama[1773]: time=2024-11-10T21:32:56.446+08:00 level=INFO source=runner.go:863 msg="starting go runner"
ollama[1773]: time=2024-11-10T21:32:56.446+08:00 level=INFO source=runner.go:864 msg=system info="AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | RISCV_VECT = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 1 | cgo(gcc)" threads=6
ollama[1773]: time=2024-11-10T21:32:56.446+08:00 level=INFO source=.:0 msg="Server listening on 127.0.0.1:40225"
ollama[1773]: llama_model_loader: loaded meta data with 27 key-value pairs and 396 tensors from /var/lib/ollama/models/blobs/sha256-11f274007f093fefeec994a5dbbb33d0733a4feb87f7ab66dcd7c1069fef0068 (version GGUF V3 (latest))
```
but from nvidia-smi nothing in there:
```
$ nvidia-smi
Sun Nov 10 21:38:22 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.03 Driver Version: 560.35.03 CUDA Version: 12.6 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA GeForce RTX 4060 Ti Off | 00000000:01:00.0 On | N/A |
| 0% 35C P8 14W / 165W | 498MiB / 16380MiB | 7% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| 0 N/A N/A 2054 G ...nim4annni-xorg-server-21.1.13/bin/X 252MiB |
| 0 N/A N/A 3315 G ...bcvgsdr9v5mjmr-picom-12.3/bin/picom 94MiB |
| 0 N/A N/A 10451 G ...irefox-132.0.1/bin/.firefox-wrapped 118MiB |
+-----------------------------------------------------------------------------------------+
```
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.4.1
|
{
"login": "kaleocheng",
"id": 7939352,
"node_id": "MDQ6VXNlcjc5MzkzNTI=",
"avatar_url": "https://avatars.githubusercontent.com/u/7939352?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/kaleocheng",
"html_url": "https://github.com/kaleocheng",
"followers_url": "https://api.github.com/users/kaleocheng/followers",
"following_url": "https://api.github.com/users/kaleocheng/following{/other_user}",
"gists_url": "https://api.github.com/users/kaleocheng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/kaleocheng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/kaleocheng/subscriptions",
"organizations_url": "https://api.github.com/users/kaleocheng/orgs",
"repos_url": "https://api.github.com/users/kaleocheng/repos",
"events_url": "https://api.github.com/users/kaleocheng/events{/privacy}",
"received_events_url": "https://api.github.com/users/kaleocheng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7597/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7597/timeline
| null |
reopened
| false
|
https://api.github.com/repos/ollama/ollama/issues/4642
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4642/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4642/comments
|
https://api.github.com/repos/ollama/ollama/issues/4642/events
|
https://github.com/ollama/ollama/pull/4642
| 2,317,404,366
|
PR_kwDOJ0Z1Ps5wko13
| 4,642
|
docs(gpu): Add workaround for nvidia GPU unavailable
|
{
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/followers",
"following_url": "https://api.github.com/users/sammcj/following{/other_user}",
"gists_url": "https://api.github.com/users/sammcj/gists{/gist_id}",
"starred_url": "https://api.github.com/users/sammcj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sammcj/subscriptions",
"organizations_url": "https://api.github.com/users/sammcj/orgs",
"repos_url": "https://api.github.com/users/sammcj/repos",
"events_url": "https://api.github.com/users/sammcj/events{/privacy}",
"received_events_url": "https://api.github.com/users/sammcj/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-05-26T02:50:12
| 2024-06-06T03:51:52
| 2024-06-06T03:51:51
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4642",
"html_url": "https://github.com/ollama/ollama/pull/4642",
"diff_url": "https://github.com/ollama/ollama/pull/4642.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4642.patch",
"merged_at": null
}
|
Docs:
- Update docs to add workaround for Nvidia GPU becoming unavailable after a period of time idle.
- Minor: Markdown formatting fixes.
I see people logging issues and asking for help on Discord for this quite often, this workaround has had good success in fixing the issue for many folk.
e.g. https://github.com/ollama/ollama/issues/4604
|
{
"login": "sammcj",
"id": 862951,
"node_id": "MDQ6VXNlcjg2Mjk1MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/862951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/sammcj",
"html_url": "https://github.com/sammcj",
"followers_url": "https://api.github.com/users/sammcj/followers",
"following_url": "https://api.github.com/users/sammcj/following{/other_user}",
"gists_url": "https://api.github.com/users/sammcj/gists{/gist_id}",
"starred_url": "https://api.github.com/users/sammcj/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/sammcj/subscriptions",
"organizations_url": "https://api.github.com/users/sammcj/orgs",
"repos_url": "https://api.github.com/users/sammcj/repos",
"events_url": "https://api.github.com/users/sammcj/events{/privacy}",
"received_events_url": "https://api.github.com/users/sammcj/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4642/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4642/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8170
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8170/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8170/comments
|
https://api.github.com/repos/ollama/ollama/issues/8170/events
|
https://github.com/ollama/ollama/issues/8170
| 2,749,841,385
|
I_kwDOJ0Z1Ps6j5z_p
| 8,170
|
ollama and with_structured_output fails for new langchain-ollama==0.2.2
|
{
"login": "nomisto",
"id": 28439912,
"node_id": "MDQ6VXNlcjI4NDM5OTEy",
"avatar_url": "https://avatars.githubusercontent.com/u/28439912?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/nomisto",
"html_url": "https://github.com/nomisto",
"followers_url": "https://api.github.com/users/nomisto/followers",
"following_url": "https://api.github.com/users/nomisto/following{/other_user}",
"gists_url": "https://api.github.com/users/nomisto/gists{/gist_id}",
"starred_url": "https://api.github.com/users/nomisto/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/nomisto/subscriptions",
"organizations_url": "https://api.github.com/users/nomisto/orgs",
"repos_url": "https://api.github.com/users/nomisto/repos",
"events_url": "https://api.github.com/users/nomisto/events{/privacy}",
"received_events_url": "https://api.github.com/users/nomisto/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 7
| 2024-12-19T10:16:24
| 2025-01-24T10:28:31
| 2024-12-20T21:45:24
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
```
pip install langchain-ollama==0.2.1 pydantic
```
and
```python
from langchain_ollama import ChatOllama
from typing import Optional
from pydantic import BaseModel, Field
class Person(BaseModel):
name: str
age: int
llm = ChatOllama(
model="llama3.1:latest",
base_url="http://10.103.251.101:11434",
temperature=0,
).with_structured_output(Person)
llm.invoke("Erick 27")
```
correctly returns `Person(name='Erick', age=27)`
`pip install langchain-ollama==0.2.2 pydantic` however returns `None` for the same code
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
ollama version is 0.3.5
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8170/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8170/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8239
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8239/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8239/comments
|
https://api.github.com/repos/ollama/ollama/issues/8239/events
|
https://github.com/ollama/ollama/issues/8239
| 2,758,733,492
|
I_kwDOJ0Z1Ps6kbu60
| 8,239
|
GPU is not being used on macOS when launching from CLI
|
{
"login": "Bhavya031",
"id": 98141026,
"node_id": "U_kgDOBdmDYg",
"avatar_url": "https://avatars.githubusercontent.com/u/98141026?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Bhavya031",
"html_url": "https://github.com/Bhavya031",
"followers_url": "https://api.github.com/users/Bhavya031/followers",
"following_url": "https://api.github.com/users/Bhavya031/following{/other_user}",
"gists_url": "https://api.github.com/users/Bhavya031/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Bhavya031/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Bhavya031/subscriptions",
"organizations_url": "https://api.github.com/users/Bhavya031/orgs",
"repos_url": "https://api.github.com/users/Bhavya031/repos",
"events_url": "https://api.github.com/users/Bhavya031/events{/privacy}",
"received_events_url": "https://api.github.com/users/Bhavya031/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 10
| 2024-12-25T11:04:31
| 2024-12-27T11:41:56
| 2024-12-27T11:41:56
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
On macOS, if you use Ollama, it utilizes the GPU. However, when launching via CLI, it does not. I searched for GPU flags but couldn’t find any. We need default GPU support for macOS when using the CLI.
https://github.com/user-attachments/assets/26fd9f8a-94f8-458f-8482-bbb96ab40697
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.5.4
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8239/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8239/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5275
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5275/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5275/comments
|
https://api.github.com/repos/ollama/ollama/issues/5275/events
|
https://github.com/ollama/ollama/issues/5275
| 2,373,009,336
|
I_kwDOJ0Z1Ps6NcT-4
| 5,275
|
ROCm on WSL
|
{
"login": "justinkb",
"id": 218024,
"node_id": "MDQ6VXNlcjIxODAyNA==",
"avatar_url": "https://avatars.githubusercontent.com/u/218024?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/justinkb",
"html_url": "https://github.com/justinkb",
"followers_url": "https://api.github.com/users/justinkb/followers",
"following_url": "https://api.github.com/users/justinkb/following{/other_user}",
"gists_url": "https://api.github.com/users/justinkb/gists{/gist_id}",
"starred_url": "https://api.github.com/users/justinkb/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/justinkb/subscriptions",
"organizations_url": "https://api.github.com/users/justinkb/orgs",
"repos_url": "https://api.github.com/users/justinkb/repos",
"events_url": "https://api.github.com/users/justinkb/events{/privacy}",
"received_events_url": "https://api.github.com/users/justinkb/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
},
{
"id": 6677675697,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgU-sQ",
"url": "https://api.github.com/repos/ollama/ollama/labels/wsl",
"name": "wsl",
"color": "7E0821",
"default": false,
"description": "Issues using WSL"
}
] |
open
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 15
| 2024-06-25T15:37:46
| 2025-01-23T23:15:06
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Recently, AMD released preview drivers for Windows that, alongside userspace packages for WSL, enable one to use ROCm through WSL. Ollama detection of AMD GPUs in linux, however, uses the presence of loaded amdgpu drivers and other sysfs stuff to determine various properties of the GPU. These are not available with this WSL ROCm setup, nor is rocm-smi used for querying VRAM size and its usage etc. I was wondering if it was feasible to add some detection for this setup, so it can be used anyway, even if some runtime information is not available. Is runtime knowledge of the available VRAM strictly necessary? Could a user just not make sure not to load too big of a model, and in case of failing to do so, accept that the ROCm runtime will hard error out on failing hipMallocs etc? Perhaps we could warn users in the output that this might happen.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5275/reactions",
"total_count": 5,
"+1": 5,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5275/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/5855
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5855/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5855/comments
|
https://api.github.com/repos/ollama/ollama/issues/5855/events
|
https://github.com/ollama/ollama/pull/5855
| 2,423,245,009
|
PR_kwDOJ0Z1Ps52HIZc
| 5,855
|
Remove no longer supported max vram var
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-07-22T16:09:01
| 2024-07-22T17:36:30
| 2024-07-22T17:35:29
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5855",
"html_url": "https://github.com/ollama/ollama/pull/5855",
"diff_url": "https://github.com/ollama/ollama/pull/5855.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5855.patch",
"merged_at": "2024-07-22T17:35:29"
}
|
The OLLAMA_MAX_VRAM env var was a temporary workaround for OOM scenarios. With Concurrency this was no longer wired up, and the simplistic value doesn't map to multi-GPU setups. Users can still set `num_gpu` to limit memory usage to avoid OOM if we get our predictions wrong.
Fixes #5754
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5855/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5855/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6408
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6408/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6408/comments
|
https://api.github.com/repos/ollama/ollama/issues/6408/events
|
https://github.com/ollama/ollama/issues/6408
| 2,472,334,391
|
I_kwDOJ0Z1Ps6TXNQ3
| 6,408
|
404 POST "/api/chat"
|
{
"login": "turndown",
"id": 57825084,
"node_id": "MDQ6VXNlcjU3ODI1MDg0",
"avatar_url": "https://avatars.githubusercontent.com/u/57825084?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/turndown",
"html_url": "https://github.com/turndown",
"followers_url": "https://api.github.com/users/turndown/followers",
"following_url": "https://api.github.com/users/turndown/following{/other_user}",
"gists_url": "https://api.github.com/users/turndown/gists{/gist_id}",
"starred_url": "https://api.github.com/users/turndown/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/turndown/subscriptions",
"organizations_url": "https://api.github.com/users/turndown/orgs",
"repos_url": "https://api.github.com/users/turndown/repos",
"events_url": "https://api.github.com/users/turndown/events{/privacy}",
"received_events_url": "https://api.github.com/users/turndown/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] |
closed
| false
| null |
[] | null | 12
| 2024-08-19T02:41:49
| 2024-11-05T11:02:35
| 2024-09-02T03:05:24
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
At first, it started running normally, but after a while, it reported 404,and can‘t run any model.
Can you help me solve it?Thx.
install by:curl -fsSL https://ollama.com/install.sh
**log below:**
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_print_meta: LF token = 148848 'ÄĬ'
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_print_meta: EOT token = 151643 '<|endoftext|>'
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_print_meta: max token length = 256
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: ggml_cuda_init: found 1 CUDA devices:
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: Device 0: NVIDIA A100-PCIE-40GB, compute capability 8.0, VMM: yes
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_tensors: ggml ctx size = 0.30 MiB
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_tensors: offloading 28 repeating layers to GPU
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_tensors: offloading non-repeating layers to GPU
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_tensors: offloaded 29/29 layers to GPU
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_tensors: CPU buffer size = 292.36 MiB
Aug 19 10:25:57 ecs-lcdsj ollama[1026502]: llm_load_tensors: CUDA0 buffer size = 3928.07 MiB
Aug 19 10:26:01 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:01 | 404 | 185.499µs | ::1 | POST "/api/chat"
Aug 19 10:26:02 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:02 | 200 | 1.273346ms | 172.17.0.2 | GET "/api/tags"
Aug 19 10:26:02 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:02 | 200 | 88.559µs | 172.17.0.2 | GET "/api/vers>
Aug 19 10:26:26 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:26 | 200 | 207.009µs | 127.0.0.1 | HEAD "/"
Aug 19 10:26:26 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:26 | 200 | 1.100698ms | 127.0.0.1 | GET "/api/tags"
Aug 19 10:26:33 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:33 | 200 | 46.933µs | 127.0.0.1 | HEAD "/"
Aug 19 10:26:33 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:26:33 | 200 | 23.522263ms | 127.0.0.1 | POST "/api/show"
Aug 19 10:26:44 ecs-lcdsj ollama[1026502]: time=2024-08-19T10:26:44.502+08:00 level=INFO source=server.go:627 msg="waiting for serve>
Aug 19 10:26:44 ecs-lcdsj ollama[1026502]: time=2024-08-19T10:26:44.780+08:00 level=INFO source=server.go:627 msg="waiting for serve>
Aug 19 10:27:01 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:27:01 | 404 | 7.051455ms | ::1 | POST "/api/chat"
Aug 19 10:28:01 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:28:01 | 404 | 367.924µs | ::1 | POST "/api/chat"
Aug 19 10:28:55 ecs-lcdsj systemd[1]: Stopping Ollama Service...
Aug 19 10:28:55 ecs-lcdsj ollama[1026502]: time=2024-08-19T10:28:55.817+08:00 level=WARN source=server.go:600 msg="client connection>
Aug 19 10:28:55 ecs-lcdsj ollama[1026502]: time=2024-08-19T10:28:55.818+08:00 level=ERROR source=sched.go:451 msg="error loading lla>
Aug 19 10:28:55 ecs-lcdsj ollama[1026502]: [GIN] 2024/08/19 - 10:28:55 | 499 | 3m0s | 172.17.0.2 | POST "/api/chat"
Aug 19 10:28:56 ecs-lcdsj systemd[1]: ollama.service: Succeeded.
Aug 19 10:28:56 ecs-lcdsj systemd[1]: Stopped Ollama Service.
Aug 19 10:28:56 ecs-lcdsj systemd[1]: Started Ollama Service.
Aug 19 10:28:56 ecs-lcdsj ollama[1032507]: 2024/08/19 10:28:56 routes.go:1125: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU>
Aug 19 10:28:56 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:28:56.246+08:00 level=INFO source=images.go:782 msg="total blobs: 15"
Aug 19 10:28:56 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:28:56.249+08:00 level=INFO source=images.go:790 msg="total unused blob>
Aug 19 10:28:56 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:28:56.249+08:00 level=INFO source=routes.go:1172 msg="Listening on [::>
Aug 19 10:28:56 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:28:56.250+08:00 level=INFO source=payload.go:30 msg="extracting embedd>
Aug 19 10:29:01 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:29:01.035+08:00 level=INFO source=payload.go:44 msg="Dynamic LLM libra>
Aug 19 10:29:01 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:29:01.037+08:00 level=INFO source=gpu.go:204 msg="looking for compatib>
Aug 19 10:29:10 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:29:10.605+08:00 level=INFO source=types.go:105 msg="inference compute">
Aug 19 10:29:10 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:29:10.606+08:00 level=INFO source=types.go:105 msg="inference compute">
Aug 19 10:29:10 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:29:10.606+08:00 level=INFO source=types.go:105 msg="inference compute">
Aug 19 10:29:10 ecs-lcdsj ollama[1032507]: time=2024-08-19T10:29:10.606+08:00 level=INFO source=types.go:105 msg="inference compute">
Aug 19 10:29:10 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:29:10 | 404 | 13.419583ms | ::1 | POST "/api/chat"
Aug 19 10:30:01 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:30:01 | 404 | 990.349µs | ::1 | POST "/api/chat"
Aug 19 10:31:01 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:31:01 | 404 | 224.61µs | ::1 | POST "/api/chat"
Aug 19 10:32:01 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:32:01 | 404 | 15.250541ms | ::1 | POST "/api/chat"
Aug 19 10:32:27 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:32:27 | 200 | 46.654µs | 127.0.0.1 | GET "/api/vers>
Aug 19 10:33:01 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:33:01 | 404 | 959.34µs | ::1 | POST "/api/chat"
Aug 19 10:34:02 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:34:02 | 404 | 18.592866ms | ::1 | POST "/api/chat"
Aug 19 10:35:01 ecs-lcdsj ollama[1032507]: [GIN] 2024/08/19 - 10:35:01 | 404 | 284.394µs | ::1 | POST "/api/chat"
### OS
Linux
### GPU
Nvidia
### CPU
Other
### Ollama version
0.3.6
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6408/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6408/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/179
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/179/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/179/comments
|
https://api.github.com/repos/ollama/ollama/issues/179/events
|
https://github.com/ollama/ollama/pull/179
| 1,816,921,968
|
PR_kwDOJ0Z1Ps5WJ69E
| 179
|
change push to chunked uploads from monolithic
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-22T23:16:24
| 2023-07-23T00:31:27
| 2023-07-23T00:31:26
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/179",
"html_url": "https://github.com/ollama/ollama/pull/179",
"diff_url": "https://github.com/ollama/ollama/pull/179.diff",
"patch_url": "https://github.com/ollama/ollama/pull/179.patch",
"merged_at": "2023-07-23T00:31:26"
}
| null |
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/179/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/179/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7254
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7254/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7254/comments
|
https://api.github.com/repos/ollama/ollama/issues/7254/events
|
https://github.com/ollama/ollama/issues/7254
| 2,597,953,873
|
I_kwDOJ0Z1Ps6a2aFR
| 7,254
|
Support directly running GGUF files without importing
|
{
"login": "ahizap",
"id": 67712951,
"node_id": "MDQ6VXNlcjY3NzEyOTUx",
"avatar_url": "https://avatars.githubusercontent.com/u/67712951?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ahizap",
"html_url": "https://github.com/ahizap",
"followers_url": "https://api.github.com/users/ahizap/followers",
"following_url": "https://api.github.com/users/ahizap/following{/other_user}",
"gists_url": "https://api.github.com/users/ahizap/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ahizap/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ahizap/subscriptions",
"organizations_url": "https://api.github.com/users/ahizap/orgs",
"repos_url": "https://api.github.com/users/ahizap/repos",
"events_url": "https://api.github.com/users/ahizap/events{/privacy}",
"received_events_url": "https://api.github.com/users/ahizap/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 1
| 2024-10-18T16:34:56
| 2024-12-20T04:40:13
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
In llama.cpp we can directly run models with `llama-cli -m your_model.gguf ` without having to import the model, It would be great if we can do the same with ollama.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7254/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7254/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/2839
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2839/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2839/comments
|
https://api.github.com/repos/ollama/ollama/issues/2839/events
|
https://github.com/ollama/ollama/issues/2839
| 2,161,701,266
|
I_kwDOJ0Z1Ps6A2PGS
| 2,839
|
keeps loading but never success
|
{
"login": "xudong2019",
"id": 16278392,
"node_id": "MDQ6VXNlcjE2Mjc4Mzky",
"avatar_url": "https://avatars.githubusercontent.com/u/16278392?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/xudong2019",
"html_url": "https://github.com/xudong2019",
"followers_url": "https://api.github.com/users/xudong2019/followers",
"following_url": "https://api.github.com/users/xudong2019/following{/other_user}",
"gists_url": "https://api.github.com/users/xudong2019/gists{/gist_id}",
"starred_url": "https://api.github.com/users/xudong2019/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/xudong2019/subscriptions",
"organizations_url": "https://api.github.com/users/xudong2019/orgs",
"repos_url": "https://api.github.com/users/xudong2019/repos",
"events_url": "https://api.github.com/users/xudong2019/events{/privacy}",
"received_events_url": "https://api.github.com/users/xudong2019/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6947643302,
"node_id": "LA_kwDOJ0Z1Ps8AAAABnhyfpg",
"url": "https://api.github.com/repos/ollama/ollama/labels/create",
"name": "create",
"color": "b60205",
"default": false,
"description": "Issues relating to ollama create"
}
] |
open
| false
| null |
[] | null | 3
| 2024-02-29T16:55:00
| 2024-11-06T18:00:06
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
ollama run renxin_query_type_classify "hello"

I successfully generate a model from gguf file. however keeps loading but never succeed... Any idea what's happening?
FROM ./model_query_type_classify.gguf
PARAMETER temperature 0
SYSTEM """
classify user type
"""
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2839/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2839/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/1474
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1474/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1474/comments
|
https://api.github.com/repos/ollama/ollama/issues/1474/events
|
https://github.com/ollama/ollama/issues/1474
| 2,036,649,047
|
I_kwDOJ0Z1Ps55ZMxX
| 1,474
|
subprocess or pexpect rather than the API
|
{
"login": "MikeyBeez",
"id": 14264000,
"node_id": "MDQ6VXNlcjE0MjY0MDAw",
"avatar_url": "https://avatars.githubusercontent.com/u/14264000?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/MikeyBeez",
"html_url": "https://github.com/MikeyBeez",
"followers_url": "https://api.github.com/users/MikeyBeez/followers",
"following_url": "https://api.github.com/users/MikeyBeez/following{/other_user}",
"gists_url": "https://api.github.com/users/MikeyBeez/gists{/gist_id}",
"starred_url": "https://api.github.com/users/MikeyBeez/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/MikeyBeez/subscriptions",
"organizations_url": "https://api.github.com/users/MikeyBeez/orgs",
"repos_url": "https://api.github.com/users/MikeyBeez/repos",
"events_url": "https://api.github.com/users/MikeyBeez/events{/privacy}",
"received_events_url": "https://api.github.com/users/MikeyBeez/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2023-12-11T22:34:32
| 2023-12-11T22:49:56
| 2023-12-11T22:46:01
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I find that Ollama is fast enough, but the API is very slow. I've been trying to use something like subprocess. The is program runs, but waiting for the output is torturously slow:
import subprocess
def run_ollama(model_name):
# Build the Ollama command
ollama_command = f"ollama run {model_name}"
# Start Ollama as a subprocess
process = subprocess.Popen(ollama_command, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True, shell=True)
# Enter the interactive loop
while True:
# Get user input for the prompt
user_input = input("Enter prompt (type 'exit' to end): ")
# Check if the user wants to exit
if user_input.lower() == 'exit':
break
# Send the user input to Ollama
process.stdin.write(user_input + '\n')
process.stdin.flush()
# Read and print the output from Ollama
output, error = process.communicate()
print("Ollama Output:", output.strip())
print("Ollama Error:", error.strip())
# Close the subprocess
process.stdin.close()
process.stdout.close()
process.stderr.close()
process.terminate()
if __name__ == "__main__":
# Get the model name from the command line arguments
import sys
if len(sys.argv) != 2:
print("Usage: python script.py <model_name>")
sys.exit(1)
model_name = sys.argv[1]
# Run Ollama with the specified model
run_ollama(model_name)
Attempts to stream the output as it is being created have failed. Even using the pexpect module fails, I believe because of the animated prompt. Is there a way to run this as a subprocess and get the results back word by word?
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1474/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1474/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/1256
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1256/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1256/comments
|
https://api.github.com/repos/ollama/ollama/issues/1256/events
|
https://github.com/ollama/ollama/pull/1256
| 2,008,391,507
|
PR_kwDOJ0Z1Ps5gPQiK
| 1,256
|
Implement tensor_split support in modelfile
|
{
"login": "Lissanro",
"id": 46057271,
"node_id": "MDQ6VXNlcjQ2MDU3Mjcx",
"avatar_url": "https://avatars.githubusercontent.com/u/46057271?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Lissanro",
"html_url": "https://github.com/Lissanro",
"followers_url": "https://api.github.com/users/Lissanro/followers",
"following_url": "https://api.github.com/users/Lissanro/following{/other_user}",
"gists_url": "https://api.github.com/users/Lissanro/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Lissanro/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Lissanro/subscriptions",
"organizations_url": "https://api.github.com/users/Lissanro/orgs",
"repos_url": "https://api.github.com/users/Lissanro/repos",
"events_url": "https://api.github.com/users/Lissanro/events{/privacy}",
"received_events_url": "https://api.github.com/users/Lissanro/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 7
| 2023-11-23T14:58:47
| 2024-04-08T17:15:18
| 2024-01-25T22:13:50
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1256",
"html_url": "https://github.com/ollama/ollama/pull/1256",
"diff_url": "https://github.com/ollama/ollama/pull/1256.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1256.patch",
"merged_at": null
}
|
This patch allows to specify a string for --tensor-split in a modelfile, for example:
PARAMETER tensor_split "25,75"
This allows to adjust VRAM allocation for each model, for example, to optimize VRAM usage on each GPU or to better accommodate models which need more memory for context on the main GPU.
|
{
"login": "Lissanro",
"id": 46057271,
"node_id": "MDQ6VXNlcjQ2MDU3Mjcx",
"avatar_url": "https://avatars.githubusercontent.com/u/46057271?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Lissanro",
"html_url": "https://github.com/Lissanro",
"followers_url": "https://api.github.com/users/Lissanro/followers",
"following_url": "https://api.github.com/users/Lissanro/following{/other_user}",
"gists_url": "https://api.github.com/users/Lissanro/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Lissanro/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Lissanro/subscriptions",
"organizations_url": "https://api.github.com/users/Lissanro/orgs",
"repos_url": "https://api.github.com/users/Lissanro/repos",
"events_url": "https://api.github.com/users/Lissanro/events{/privacy}",
"received_events_url": "https://api.github.com/users/Lissanro/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1256/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1256/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2307
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2307/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2307/comments
|
https://api.github.com/repos/ollama/ollama/issues/2307/events
|
https://github.com/ollama/ollama/pull/2307
| 2,112,042,563
|
PR_kwDOJ0Z1Ps5lrOJr
| 2,307
|
Fix help string for stop parameter
|
{
"login": "gaardhus",
"id": 46934916,
"node_id": "MDQ6VXNlcjQ2OTM0OTE2",
"avatar_url": "https://avatars.githubusercontent.com/u/46934916?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/gaardhus",
"html_url": "https://github.com/gaardhus",
"followers_url": "https://api.github.com/users/gaardhus/followers",
"following_url": "https://api.github.com/users/gaardhus/following{/other_user}",
"gists_url": "https://api.github.com/users/gaardhus/gists{/gist_id}",
"starred_url": "https://api.github.com/users/gaardhus/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/gaardhus/subscriptions",
"organizations_url": "https://api.github.com/users/gaardhus/orgs",
"repos_url": "https://api.github.com/users/gaardhus/repos",
"events_url": "https://api.github.com/users/gaardhus/events{/privacy}",
"received_events_url": "https://api.github.com/users/gaardhus/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-02-01T09:47:24
| 2024-05-07T23:48:35
| 2024-05-07T23:48:35
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2307",
"html_url": "https://github.com/ollama/ollama/pull/2307",
"diff_url": "https://github.com/ollama/ollama/pull/2307.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2307.patch",
"merged_at": "2024-05-07T23:48:35"
}
|
Changed the help prompt for setting the stop parameters, and quotes or commas are otherwise included in the stop-token:
/set parameter stop "?", "!" # Invalid
/set parameter stop ? ! # Valid
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2307/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2307/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4368
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4368/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4368/comments
|
https://api.github.com/repos/ollama/ollama/issues/4368/events
|
https://github.com/ollama/ollama/pull/4368
| 2,291,084,965
|
PR_kwDOJ0Z1Ps5vKx0b
| 4,368
|
Fix OpenAI `finish_reason` values when empty
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-05-11T22:31:27
| 2024-05-11T22:31:42
| 2024-05-11T22:31:41
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4368",
"html_url": "https://github.com/ollama/ollama/pull/4368",
"diff_url": "https://github.com/ollama/ollama/pull/4368.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4368.patch",
"merged_at": "2024-05-11T22:31:41"
}
|
Fixes https://github.com/ollama/ollama/issues/4357
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4368/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4368/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4127
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4127/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4127/comments
|
https://api.github.com/repos/ollama/ollama/issues/4127/events
|
https://github.com/ollama/ollama/issues/4127
| 2,277,792,321
|
I_kwDOJ0Z1Ps6HxFpB
| 4,127
|
Add LLAVA++ model
|
{
"login": "ddpasa",
"id": 112642920,
"node_id": "U_kgDOBrbLaA",
"avatar_url": "https://avatars.githubusercontent.com/u/112642920?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ddpasa",
"html_url": "https://github.com/ddpasa",
"followers_url": "https://api.github.com/users/ddpasa/followers",
"following_url": "https://api.github.com/users/ddpasa/following{/other_user}",
"gists_url": "https://api.github.com/users/ddpasa/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ddpasa/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ddpasa/subscriptions",
"organizations_url": "https://api.github.com/users/ddpasa/orgs",
"repos_url": "https://api.github.com/users/ddpasa/repos",
"events_url": "https://api.github.com/users/ddpasa/events{/privacy}",
"received_events_url": "https://api.github.com/users/ddpasa/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-05-03T14:24:21
| 2024-05-21T21:48:43
| 2024-05-21T21:48:43
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
There is a new version of the Amazing LLava model that uses Llama 3 or Phi-3:
https://huggingface.co/collections/MBZUAI/llava-llama-3-and-phi-3-mini-662b38b972e3e3e4d8f821bb
https://github.com/mbzuai-oryx/LLaVA-pp
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4127/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4127/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7897
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7897/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7897/comments
|
https://api.github.com/repos/ollama/ollama/issues/7897/events
|
https://github.com/ollama/ollama/issues/7897
| 2,707,771,502
|
I_kwDOJ0Z1Ps6hZVBu
| 7,897
|
Audio to audio models
|
{
"login": "mohammadaminyza",
"id": 73334272,
"node_id": "MDQ6VXNlcjczMzM0Mjcy",
"avatar_url": "https://avatars.githubusercontent.com/u/73334272?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mohammadaminyza",
"html_url": "https://github.com/mohammadaminyza",
"followers_url": "https://api.github.com/users/mohammadaminyza/followers",
"following_url": "https://api.github.com/users/mohammadaminyza/following{/other_user}",
"gists_url": "https://api.github.com/users/mohammadaminyza/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mohammadaminyza/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mohammadaminyza/subscriptions",
"organizations_url": "https://api.github.com/users/mohammadaminyza/orgs",
"repos_url": "https://api.github.com/users/mohammadaminyza/repos",
"events_url": "https://api.github.com/users/mohammadaminyza/events{/privacy}",
"received_events_url": "https://api.github.com/users/mohammadaminyza/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-11-30T18:31:44
| 2024-11-30T18:31:44
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi, any plan to add audio to audio support? There are couple of open source model witch provide that
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7897/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7897/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/6729
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6729/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6729/comments
|
https://api.github.com/repos/ollama/ollama/issues/6729/events
|
https://github.com/ollama/ollama/pull/6729
| 2,516,631,081
|
PR_kwDOJ0Z1Ps56_i00
| 6,729
|
Feature: Add Support for Distributed Inferencing
|
{
"login": "ecyht2",
"id": 94816144,
"node_id": "U_kgDOBabHkA",
"avatar_url": "https://avatars.githubusercontent.com/u/94816144?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ecyht2",
"html_url": "https://github.com/ecyht2",
"followers_url": "https://api.github.com/users/ecyht2/followers",
"following_url": "https://api.github.com/users/ecyht2/following{/other_user}",
"gists_url": "https://api.github.com/users/ecyht2/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ecyht2/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ecyht2/subscriptions",
"organizations_url": "https://api.github.com/users/ecyht2/orgs",
"repos_url": "https://api.github.com/users/ecyht2/repos",
"events_url": "https://api.github.com/users/ecyht2/events{/privacy}",
"received_events_url": "https://api.github.com/users/ecyht2/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 20
| 2024-09-10T14:24:43
| 2025-01-24T23:15:45
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | true
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/6729",
"html_url": "https://github.com/ollama/ollama/pull/6729",
"diff_url": "https://github.com/ollama/ollama/pull/6729.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6729.patch",
"merged_at": null
}
|
This feature adds support for llama.cpp RPC. This allows for distributed inferencing on different devices.
This Pull Request aims to implement #4643.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6729/reactions",
"total_count": 47,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 27,
"rocket": 20,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6729/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1444
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1444/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1444/comments
|
https://api.github.com/repos/ollama/ollama/issues/1444/events
|
https://github.com/ollama/ollama/pull/1444
| 2,033,527,333
|
PR_kwDOJ0Z1Ps5hkgjb
| 1,444
|
Added mention of the NOPRUNE env var
|
{
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.github.com/users/technovangelist/followers",
"following_url": "https://api.github.com/users/technovangelist/following{/other_user}",
"gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}",
"starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions",
"organizations_url": "https://api.github.com/users/technovangelist/orgs",
"repos_url": "https://api.github.com/users/technovangelist/repos",
"events_url": "https://api.github.com/users/technovangelist/events{/privacy}",
"received_events_url": "https://api.github.com/users/technovangelist/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-12-09T01:38:51
| 2023-12-12T01:15:00
| 2023-12-12T01:15:00
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1444",
"html_url": "https://github.com/ollama/ollama/pull/1444",
"diff_url": "https://github.com/ollama/ollama/pull/1444.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1444.patch",
"merged_at": null
}
|
OLLAMA_NOPRUNE will prevent the pruning process from running, but it isn't mentioned anywhere outside of the code and a merged PR.
|
{
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.github.com/users/technovangelist/followers",
"following_url": "https://api.github.com/users/technovangelist/following{/other_user}",
"gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}",
"starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions",
"organizations_url": "https://api.github.com/users/technovangelist/orgs",
"repos_url": "https://api.github.com/users/technovangelist/repos",
"events_url": "https://api.github.com/users/technovangelist/events{/privacy}",
"received_events_url": "https://api.github.com/users/technovangelist/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1444/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1444/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1027
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1027/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1027/comments
|
https://api.github.com/repos/ollama/ollama/issues/1027/events
|
https://github.com/ollama/ollama/issues/1027
| 1,980,781,895
|
I_kwDOJ0Z1Ps52EFVH
| 1,027
|
How to properly format Advanced Parameters / options in API calls?
|
{
"login": "tob-har",
"id": 32613633,
"node_id": "MDQ6VXNlcjMyNjEzNjMz",
"avatar_url": "https://avatars.githubusercontent.com/u/32613633?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/tob-har",
"html_url": "https://github.com/tob-har",
"followers_url": "https://api.github.com/users/tob-har/followers",
"following_url": "https://api.github.com/users/tob-har/following{/other_user}",
"gists_url": "https://api.github.com/users/tob-har/gists{/gist_id}",
"starred_url": "https://api.github.com/users/tob-har/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/tob-har/subscriptions",
"organizations_url": "https://api.github.com/users/tob-har/orgs",
"repos_url": "https://api.github.com/users/tob-har/repos",
"events_url": "https://api.github.com/users/tob-har/events{/privacy}",
"received_events_url": "https://api.github.com/users/tob-har/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 3
| 2023-11-07T08:10:16
| 2023-11-09T00:44:38
| 2023-11-09T00:44:38
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
API Documentation gives a proper example, how to use
`POST /api/generate`
But how to properly format the JSON object to use Advanced Parameters?
Especially `options` and `system`.
I tried to request the following via `POST /api/generate`.
Everything behaves as expected, eg stream, but options is not workig:
`
{
"model": "llama2:latest",
"stream": false,
"prompt": "Sing a song.",
"options": {
"temperature": 5}
}
`
Happy about hints!!! Thanks a lot
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1027/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
}
|
https://api.github.com/repos/ollama/ollama/issues/1027/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4392
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4392/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4392/comments
|
https://api.github.com/repos/ollama/ollama/issues/4392/events
|
https://github.com/ollama/ollama/issues/4392
| 2,292,163,129
|
I_kwDOJ0Z1Ps6In6I5
| 4,392
|
Use GTT memory in case of iGPUs to run the model efiiciently.
|
{
"login": "CoolnsX",
"id": 76195824,
"node_id": "MDQ6VXNlcjc2MTk1ODI0",
"avatar_url": "https://avatars.githubusercontent.com/u/76195824?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/CoolnsX",
"html_url": "https://github.com/CoolnsX",
"followers_url": "https://api.github.com/users/CoolnsX/followers",
"following_url": "https://api.github.com/users/CoolnsX/following{/other_user}",
"gists_url": "https://api.github.com/users/CoolnsX/gists{/gist_id}",
"starred_url": "https://api.github.com/users/CoolnsX/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/CoolnsX/subscriptions",
"organizations_url": "https://api.github.com/users/CoolnsX/orgs",
"repos_url": "https://api.github.com/users/CoolnsX/repos",
"events_url": "https://api.github.com/users/CoolnsX/events{/privacy}",
"received_events_url": "https://api.github.com/users/CoolnsX/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 1
| 2024-05-13T08:28:01
| 2024-11-02T18:48:53
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Models run on System memory using CPU is perfectly fine.
But when using integrate GPUs which have limited VRAM locked by vendors, we have model crash due to "low vram memory"
They have feature called GTT memory on linux, and Shared Memory on windows, which they can use whenever their VRAM capacity is nearly full.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4392/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4392/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/7457
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7457/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7457/comments
|
https://api.github.com/repos/ollama/ollama/issues/7457/events
|
https://github.com/ollama/ollama/issues/7457
| 2,627,853,133
|
I_kwDOJ0Z1Ps6codtN
| 7,457
|
Adding avx2+avx512 to cuda runner in new ollama code
|
{
"login": "AncientMystic",
"id": 62780271,
"node_id": "MDQ6VXNlcjYyNzgwMjcx",
"avatar_url": "https://avatars.githubusercontent.com/u/62780271?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/AncientMystic",
"html_url": "https://github.com/AncientMystic",
"followers_url": "https://api.github.com/users/AncientMystic/followers",
"following_url": "https://api.github.com/users/AncientMystic/following{/other_user}",
"gists_url": "https://api.github.com/users/AncientMystic/gists{/gist_id}",
"starred_url": "https://api.github.com/users/AncientMystic/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/AncientMystic/subscriptions",
"organizations_url": "https://api.github.com/users/AncientMystic/orgs",
"repos_url": "https://api.github.com/users/AncientMystic/repos",
"events_url": "https://api.github.com/users/AncientMystic/events{/privacy}",
"received_events_url": "https://api.github.com/users/AncientMystic/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 0
| 2024-10-31T21:19:35
| 2024-12-10T17:47:22
| 2024-12-10T17:47:22
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
In the old code i added avx2+512 in the gen_windows.ps1 by simply adding DGGML_AVX2=on & DGGML_AVX512=on after the DGGML_AVX=on line in the cuda build function
It added a fairly decent performance boost
I have added avx512 to cpu, But In the new code i cannot seem to find where to properly add it in the make files to also add it to cuda
What do i need to change on what lines in which files to add this to the new ollama code?
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7457/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7457/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3146
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3146/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3146/comments
|
https://api.github.com/repos/ollama/ollama/issues/3146/events
|
https://github.com/ollama/ollama/pull/3146
| 2,187,016,061
|
PR_kwDOJ0Z1Ps5pqosz
| 3,146
|
server: replace blob prefix separator from ':' to '-'
|
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers",
"following_url": "https://api.github.com/users/bmizerany/following{/other_user}",
"gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions",
"organizations_url": "https://api.github.com/users/bmizerany/orgs",
"repos_url": "https://api.github.com/users/bmizerany/repos",
"events_url": "https://api.github.com/users/bmizerany/events{/privacy}",
"received_events_url": "https://api.github.com/users/bmizerany/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-03-14T18:32:54
| 2024-03-25T16:22:07
| 2024-03-15T03:18:06
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3146",
"html_url": "https://github.com/ollama/ollama/pull/3146",
"diff_url": "https://github.com/ollama/ollama/pull/3146.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3146.patch",
"merged_at": "2024-03-15T03:18:06"
}
|
This fixes issues with blob file names that contain ':' characters to be
9 rejected by file systems that do not support them.
|
{
"login": "bmizerany",
"id": 46,
"node_id": "MDQ6VXNlcjQ2",
"avatar_url": "https://avatars.githubusercontent.com/u/46?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/bmizerany",
"html_url": "https://github.com/bmizerany",
"followers_url": "https://api.github.com/users/bmizerany/followers",
"following_url": "https://api.github.com/users/bmizerany/following{/other_user}",
"gists_url": "https://api.github.com/users/bmizerany/gists{/gist_id}",
"starred_url": "https://api.github.com/users/bmizerany/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/bmizerany/subscriptions",
"organizations_url": "https://api.github.com/users/bmizerany/orgs",
"repos_url": "https://api.github.com/users/bmizerany/repos",
"events_url": "https://api.github.com/users/bmizerany/events{/privacy}",
"received_events_url": "https://api.github.com/users/bmizerany/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3146/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3146/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6653
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6653/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6653/comments
|
https://api.github.com/repos/ollama/ollama/issues/6653/events
|
https://github.com/ollama/ollama/issues/6653
| 2,507,275,752
|
I_kwDOJ0Z1Ps6Vcf3o
| 6,653
|
Loading a smaller context model after a bigger model is loaded
|
{
"login": "Madhav-Gohel",
"id": 76510494,
"node_id": "MDQ6VXNlcjc2NTEwNDk0",
"avatar_url": "https://avatars.githubusercontent.com/u/76510494?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Madhav-Gohel",
"html_url": "https://github.com/Madhav-Gohel",
"followers_url": "https://api.github.com/users/Madhav-Gohel/followers",
"following_url": "https://api.github.com/users/Madhav-Gohel/following{/other_user}",
"gists_url": "https://api.github.com/users/Madhav-Gohel/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Madhav-Gohel/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Madhav-Gohel/subscriptions",
"organizations_url": "https://api.github.com/users/Madhav-Gohel/orgs",
"repos_url": "https://api.github.com/users/Madhav-Gohel/repos",
"events_url": "https://api.github.com/users/Madhav-Gohel/events{/privacy}",
"received_events_url": "https://api.github.com/users/Madhav-Gohel/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 0
| 2024-09-05T09:32:46
| 2024-09-05T09:32:46
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
## Hardware
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 46 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 64
On-line CPU(s) list: 0-63
Vendor ID: GenuineIntel
Model name: Intel(R) Xeon(R) Gold 5218 CPU @ 2.30GHz
CPU family: 6
Model: 85
Thread(s) per core: 2
Core(s) per socket: 16
Socket(s): 2
3 x Quadro RTX 5000

## Error
The below happens when **llama3.1** is already loaded and i am loading **smollm** which are having different context length.
Both model are loaded into gpu on request but results below on API request

Both model works fine concurrently using ollama cli
### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.3.5
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6653/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6653/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/2037
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2037/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2037/comments
|
https://api.github.com/repos/ollama/ollama/issues/2037/events
|
https://github.com/ollama/ollama/pull/2037
| 2,087,274,232
|
PR_kwDOJ0Z1Ps5kXvPO
| 2,037
|
fix: pasting slash commands
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2024-01-18T01:00:22
| 2025-01-15T02:54:49
| 2025-01-15T02:54:44
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2037",
"html_url": "https://github.com/ollama/ollama/pull/2037",
"diff_url": "https://github.com/ollama/ollama/pull/2037.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2037.patch",
"merged_at": null
}
|
there is a bug in paste where the pasted content is written directly to the prompt buffer instead of being processed. for most content, this is fine but slash commands are processed line-by-line.
aggregate status updates, e.g. "Set 'verbose' mode.", "Set system message.", to the end for aesthetics. the status message shouldn't display while in paste mode
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2037/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2037/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8050
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8050/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8050/comments
|
https://api.github.com/repos/ollama/ollama/issues/8050/events
|
https://github.com/ollama/ollama/issues/8050
| 2,733,416,718
|
I_kwDOJ0Z1Ps6i7KEO
| 8,050
|
Ollama behind proxy can't pull new models anymore
|
{
"login": "the-silversurver",
"id": 135591792,
"node_id": "U_kgDOCBT3cA",
"avatar_url": "https://avatars.githubusercontent.com/u/135591792?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/the-silversurver",
"html_url": "https://github.com/the-silversurver",
"followers_url": "https://api.github.com/users/the-silversurver/followers",
"following_url": "https://api.github.com/users/the-silversurver/following{/other_user}",
"gists_url": "https://api.github.com/users/the-silversurver/gists{/gist_id}",
"starred_url": "https://api.github.com/users/the-silversurver/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/the-silversurver/subscriptions",
"organizations_url": "https://api.github.com/users/the-silversurver/orgs",
"repos_url": "https://api.github.com/users/the-silversurver/repos",
"events_url": "https://api.github.com/users/the-silversurver/events{/privacy}",
"received_events_url": "https://api.github.com/users/the-silversurver/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
},
{
"id": 6677370291,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw",
"url": "https://api.github.com/repos/ollama/ollama/labels/networking",
"name": "networking",
"color": "0B5368",
"default": false,
"description": "Issues relating to ollama pull and push"
}
] |
closed
| false
| null |
[] | null | 12
| 2024-12-11T16:21:38
| 2025-01-13T01:38:03
| 2025-01-13T01:38:03
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hi there,
I am using Ollama on different machines (Ubuntu inside a docker container together with open web ui and on a Mac standalone) within a university that enforces the use of a proxy to access the internet.
On both systems, the proxy is correctly configured and Ollama worked with it, be it via zsh
ollama pull modelname
or
via open web ui.
Now on both systems pulling new models does not work anymore.
The Ollama container at the ubuntu server says:
`time=2024-12-11T14:34:35.279Z level=INFO source=images.go:990 msg="request failed: Get \"http://registry.ollama.ai/v2/library/qwen2.5-coder/manifests/latest\": dial tcp 104.21.75.227:80: connect: connection refused"
[GIN] 2024/12/11 - 14:34:35 | 200 | 55.03224ms | 172.31.1.5 | POST "/api/pull"
`
On the Mac I get
`Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/qwen2.5-coder/manifests/latest": dial tcp 172.67.182.229:443: connect: connection refused`
On both systems
`wget https://registry.ollama.ai/v2/library/qwen2.5-coder/manifests/latest` pulls the manifest correctly.
When I deactivate the LAN connection on the Mac and use WLAN to connect directly to the internet without the proxy, anything works as expected.
This behavior is new to me, as I am running ollama for several months on both machines and never had these issues.
Any hint is highly appreciated
for the Ollama - Ollama web ui on the ubuntu server this is my docker-compose.yml which worked fine in the past
`services:
ollama:
deploy:
resources:
reservations:
devices:
- driver: ${OLLAMA_GPU_DRIVER-nvidia}
count: all
capabilities:
- gpu
volumes:
- /docker-apps/ollama-web-ui/volumes/ollama:/root/.ollama
container_name: ollama
pull_policy: always
tty: true
restart: unless-stopped
image: ollama/ollama:${OLLAMA_DOCKER_TAG-latest}
environment:
- HTTP_PROXY=http://proxy.my-uni.tld:8080
- HTTPS_PROXY=http://proxy.my-uni.tld:8080
- NO_PROXY=localhost,127.0.0.1,.my-uni.tld
ports:
- 11434:11434
networks:
- ollama_network
open-webui:
build:
context: .
args:
OLLAMA_BASE_URL: '/ollama'
dockerfile: Dockerfile
image: ghcr.io/open-webui/open-webui:${WEBUI_DOCKER_TAG-main}
container_name: open-webui
volumes:
- /docker-apps/ollama-web-ui/volumes/open-web-ui:/app/backend/data
depends_on:
- ollama
networks:
- ollama_network
environment:
- 'OLLAMA_BASE_URL=https://quest-gpu-06.my-uni.tld'
- 'WEBUI_SECRET_KEY=xyz'
- HTTP_PROXY=http://proxy.my-uni.tld:8080
- HTTPS_PROXY=http://proxy.my-uni.tld:8080
- NO_PROXY=localhost,127.0.0.1,.my-uni.tld
extra_hosts:
- host.docker.internal:host-gateway
restart: unless-stopped
ports:
- 8080:8080
networks:
ollama_network:
external: true`
### OS
Linux, macOS
### GPU
Nvidia, Apple
### CPU
AMD, Apple
### Ollama version
0.4.4
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8050/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8050/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5134
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5134/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5134/comments
|
https://api.github.com/repos/ollama/ollama/issues/5134/events
|
https://github.com/ollama/ollama/issues/5134
| 2,361,410,869
|
I_kwDOJ0Z1Ps6MwEU1
| 5,134
|
api interface /api/generate I need to make sure that every question is not answered from the previous record How to do?
|
{
"login": "mingLvft",
"id": 50644675,
"node_id": "MDQ6VXNlcjUwNjQ0Njc1",
"avatar_url": "https://avatars.githubusercontent.com/u/50644675?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mingLvft",
"html_url": "https://github.com/mingLvft",
"followers_url": "https://api.github.com/users/mingLvft/followers",
"following_url": "https://api.github.com/users/mingLvft/following{/other_user}",
"gists_url": "https://api.github.com/users/mingLvft/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mingLvft/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mingLvft/subscriptions",
"organizations_url": "https://api.github.com/users/mingLvft/orgs",
"repos_url": "https://api.github.com/users/mingLvft/repos",
"events_url": "https://api.github.com/users/mingLvft/events{/privacy}",
"received_events_url": "https://api.github.com/users/mingLvft/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-06-19T06:09:53
| 2024-11-20T20:11:36
| 2024-06-27T21:33:43
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
api interface /api/generate I need to make sure that every question is not answered from the previous record How to do?
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5134/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5134/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/7810
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7810/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7810/comments
|
https://api.github.com/repos/ollama/ollama/issues/7810/events
|
https://github.com/ollama/ollama/issues/7810
| 2,685,926,172
|
I_kwDOJ0Z1Ps6gF_sc
| 7,810
|
could anyone help me? something is not work. use a special gpu
|
{
"login": "wangzd0209",
"id": 99313728,
"node_id": "U_kgDOBetoQA",
"avatar_url": "https://avatars.githubusercontent.com/u/99313728?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wangzd0209",
"html_url": "https://github.com/wangzd0209",
"followers_url": "https://api.github.com/users/wangzd0209/followers",
"following_url": "https://api.github.com/users/wangzd0209/following{/other_user}",
"gists_url": "https://api.github.com/users/wangzd0209/gists{/gist_id}",
"starred_url": "https://api.github.com/users/wangzd0209/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/wangzd0209/subscriptions",
"organizations_url": "https://api.github.com/users/wangzd0209/orgs",
"repos_url": "https://api.github.com/users/wangzd0209/repos",
"events_url": "https://api.github.com/users/wangzd0209/events{/privacy}",
"received_events_url": "https://api.github.com/users/wangzd0209/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 1
| 2024-11-23T12:07:01
| 2024-12-01T02:19:24
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
when i follow the instruction to install ollama with source code. i can not to finish gen.linux.sh
there are error information
`CMake Error at ggml/src/CMakeLists.txt:440 (find_package):
By not providing "Findhip.cmake" in CMAKE_MODULE_PATH this project has
asked CMake to find a package configuration file provided by "hip", but
CMake did not find one.
Could not find a package configuration file provided by "hip" with any of
the following names:
hipConfig.cmake
hip-config.cmake
Add the installation prefix of "hip" to CMAKE_PREFIX_PATH or set "hip_DIR"
to a directory containing one of the above files. If "hip" provides a
separate development package or SDK, be sure it has been installed.`
i try to modify as [https://github.com/ROCm/HIP/tree/master/samples/2_Cookbook/12_cmake_hip_add_executable#including-findhip-cmake-module-in-the-project](url)
but the cmakelist can not be modify, when i restart gen.linux.sh , camkelists.txt on ggml/src do not modify
could anyone help me?
### OS
Linux
### GPU
AMD, Other
### CPU
_No response_
### Ollama version
_No response_
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7810/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7810/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/3112
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3112/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3112/comments
|
https://api.github.com/repos/ollama/ollama/issues/3112/events
|
https://github.com/ollama/ollama/issues/3112
| 2,184,294,406
|
I_kwDOJ0Z1Ps6CMbAG
| 3,112
|
Windows Error:pull model manifest return wsarecv: An existing connection was forcibly closed by the remote host.
|
{
"login": "heimu-liu",
"id": 102661308,
"node_id": "U_kgDOBh58vA",
"avatar_url": "https://avatars.githubusercontent.com/u/102661308?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/heimu-liu",
"html_url": "https://github.com/heimu-liu",
"followers_url": "https://api.github.com/users/heimu-liu/followers",
"following_url": "https://api.github.com/users/heimu-liu/following{/other_user}",
"gists_url": "https://api.github.com/users/heimu-liu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/heimu-liu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/heimu-liu/subscriptions",
"organizations_url": "https://api.github.com/users/heimu-liu/orgs",
"repos_url": "https://api.github.com/users/heimu-liu/repos",
"events_url": "https://api.github.com/users/heimu-liu/events{/privacy}",
"received_events_url": "https://api.github.com/users/heimu-liu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
},
{
"id": 6677370291,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw",
"url": "https://api.github.com/repos/ollama/ollama/labels/networking",
"name": "networking",
"color": "0B5368",
"default": false,
"description": "Issues relating to ollama pull and push"
}
] |
closed
| false
| null |
[] | null | 11
| 2024-03-13T15:26:36
| 2024-04-23T05:27:48
| 2024-03-29T03:25:30
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
i can't down the model:
[app.log](https://github.com/ollama/ollama/files/14589977/app.log)
[server.log](https://github.com/ollama/ollama/files/14589978/server.log)
`PS C:\Users\heimu\AppData\Local\Ollama> ollama pull llama2
pulling manifest
Error: pull model manifest: Get "https://ollama.com/token?nonce=A-QmGZFS0za-Kv0GKrDy3Q&scope=repository%!A(MISSING)library%!F(MISSING)llama2%!A(MISSING)pull&service=ollama.com&ts=1710339130": read tcp 192.168.247.214:56798->34.120.132.20:443: wsarecv: An existing connection was forcibly closed by the remote host.
PS C:\Users\heimu\AppData\Local\Ollama>`
|
{
"login": "heimu-liu",
"id": 102661308,
"node_id": "U_kgDOBh58vA",
"avatar_url": "https://avatars.githubusercontent.com/u/102661308?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/heimu-liu",
"html_url": "https://github.com/heimu-liu",
"followers_url": "https://api.github.com/users/heimu-liu/followers",
"following_url": "https://api.github.com/users/heimu-liu/following{/other_user}",
"gists_url": "https://api.github.com/users/heimu-liu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/heimu-liu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/heimu-liu/subscriptions",
"organizations_url": "https://api.github.com/users/heimu-liu/orgs",
"repos_url": "https://api.github.com/users/heimu-liu/repos",
"events_url": "https://api.github.com/users/heimu-liu/events{/privacy}",
"received_events_url": "https://api.github.com/users/heimu-liu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3112/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3112/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2061
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2061/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2061/comments
|
https://api.github.com/repos/ollama/ollama/issues/2061/events
|
https://github.com/ollama/ollama/pull/2061
| 2,089,359,047
|
PR_kwDOJ0Z1Ps5ke4Nm
| 2,061
|
ci: use stubs libraries
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-01-19T00:55:16
| 2024-01-19T01:17:47
| 2024-01-19T01:17:44
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | true
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2061",
"html_url": "https://github.com/ollama/ollama/pull/2061",
"diff_url": "https://github.com/ollama/ollama/pull/2061.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2061.patch",
"merged_at": null
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2061/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2061/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1523
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1523/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1523/comments
|
https://api.github.com/repos/ollama/ollama/issues/1523/events
|
https://github.com/ollama/ollama/issues/1523
| 2,041,859,796
|
I_kwDOJ0Z1Ps55tE7U
| 1,523
|
docs: generate chat response `loadDuration` missing
|
{
"login": "mthongvanh",
"id": 4961248,
"node_id": "MDQ6VXNlcjQ5NjEyNDg=",
"avatar_url": "https://avatars.githubusercontent.com/u/4961248?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mthongvanh",
"html_url": "https://github.com/mthongvanh",
"followers_url": "https://api.github.com/users/mthongvanh/followers",
"following_url": "https://api.github.com/users/mthongvanh/following{/other_user}",
"gists_url": "https://api.github.com/users/mthongvanh/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mthongvanh/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mthongvanh/subscriptions",
"organizations_url": "https://api.github.com/users/mthongvanh/orgs",
"repos_url": "https://api.github.com/users/mthongvanh/repos",
"events_url": "https://api.github.com/users/mthongvanh/events{/privacy}",
"received_events_url": "https://api.github.com/users/mthongvanh/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1
| 2023-12-14T14:49:34
| 2023-12-14T17:15:51
| 2023-12-14T17:15:51
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
in the documentation https://github.com/jmorganca/ollama/blob/main/docs/api.md#response-6 loadDuration is listed as a return value but does not get returned by the api
<img width="981" alt="image" src="https://github.com/jmorganca/ollama/assets/4961248/bb0dbc37-c2cf-48ff-8c8d-be2ffcfa5115">
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1523/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1523/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/103
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/103/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/103/comments
|
https://api.github.com/repos/ollama/ollama/issues/103/events
|
https://github.com/ollama/ollama/pull/103
| 1,810,624,762
|
PR_kwDOJ0Z1Ps5V0lpv
| 103
|
website content and design update
|
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2023-07-18T19:58:33
| 2023-07-23T10:25:30
| 2023-07-18T20:18:05
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/103",
"html_url": "https://github.com/ollama/ollama/pull/103",
"diff_url": "https://github.com/ollama/ollama/pull/103.diff",
"patch_url": "https://github.com/ollama/ollama/pull/103.patch",
"merged_at": "2023-07-18T20:18:04"
}
| null |
{
"login": "hoyyeva",
"id": 63033505,
"node_id": "MDQ6VXNlcjYzMDMzNTA1",
"avatar_url": "https://avatars.githubusercontent.com/u/63033505?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/hoyyeva",
"html_url": "https://github.com/hoyyeva",
"followers_url": "https://api.github.com/users/hoyyeva/followers",
"following_url": "https://api.github.com/users/hoyyeva/following{/other_user}",
"gists_url": "https://api.github.com/users/hoyyeva/gists{/gist_id}",
"starred_url": "https://api.github.com/users/hoyyeva/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/hoyyeva/subscriptions",
"organizations_url": "https://api.github.com/users/hoyyeva/orgs",
"repos_url": "https://api.github.com/users/hoyyeva/repos",
"events_url": "https://api.github.com/users/hoyyeva/events{/privacy}",
"received_events_url": "https://api.github.com/users/hoyyeva/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/103/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/103/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5681
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5681/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5681/comments
|
https://api.github.com/repos/ollama/ollama/issues/5681/events
|
https://github.com/ollama/ollama/pull/5681
| 2,407,157,044
|
PR_kwDOJ0Z1Ps51Tgif
| 5,681
|
Adding instructions when user doesn't have sudo privileges
|
{
"login": "Ivanknmk",
"id": 1672248,
"node_id": "MDQ6VXNlcjE2NzIyNDg=",
"avatar_url": "https://avatars.githubusercontent.com/u/1672248?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Ivanknmk",
"html_url": "https://github.com/Ivanknmk",
"followers_url": "https://api.github.com/users/Ivanknmk/followers",
"following_url": "https://api.github.com/users/Ivanknmk/following{/other_user}",
"gists_url": "https://api.github.com/users/Ivanknmk/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Ivanknmk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Ivanknmk/subscriptions",
"organizations_url": "https://api.github.com/users/Ivanknmk/orgs",
"repos_url": "https://api.github.com/users/Ivanknmk/repos",
"events_url": "https://api.github.com/users/Ivanknmk/events{/privacy}",
"received_events_url": "https://api.github.com/users/Ivanknmk/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-07-13T20:38:44
| 2024-11-25T00:02:00
| 2024-11-25T00:02:00
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5681",
"html_url": "https://github.com/ollama/ollama/pull/5681",
"diff_url": "https://github.com/ollama/ollama/pull/5681.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5681.patch",
"merged_at": null
}
|
Adding instructions when user doesn't have sudo privileges according to https://github.com/ollama/ollama/issues/2111
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5681/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5681/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/6287
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6287/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6287/comments
|
https://api.github.com/repos/ollama/ollama/issues/6287/events
|
https://github.com/ollama/ollama/issues/6287
| 2,458,234,863
|
I_kwDOJ0Z1Ps6Sha_v
| 6,287
|
UHD intel GPU Accelerate
|
{
"login": "jomardyan",
"id": 18527406,
"node_id": "MDQ6VXNlcjE4NTI3NDA2",
"avatar_url": "https://avatars.githubusercontent.com/u/18527406?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jomardyan",
"html_url": "https://github.com/jomardyan",
"followers_url": "https://api.github.com/users/jomardyan/followers",
"following_url": "https://api.github.com/users/jomardyan/following{/other_user}",
"gists_url": "https://api.github.com/users/jomardyan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jomardyan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jomardyan/subscriptions",
"organizations_url": "https://api.github.com/users/jomardyan/orgs",
"repos_url": "https://api.github.com/users/jomardyan/repos",
"events_url": "https://api.github.com/users/jomardyan/events{/privacy}",
"received_events_url": "https://api.github.com/users/jomardyan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6677491450,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgJu-g",
"url": "https://api.github.com/repos/ollama/ollama/labels/intel",
"name": "intel",
"color": "226E5B",
"default": false,
"description": "issues relating to Intel GPUs"
}
] |
closed
| false
| null |
[] | null | 2
| 2024-08-09T16:03:25
| 2024-08-28T02:52:37
| 2024-08-09T18:36:38
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Why Ollama use CPU, but not utilizing intel UHD integrated GPU ?
(Computer with not Nvidia GPU)
### OS
Linux
### GPU
Intel
### CPU
Intel
### Ollama version
_No response_
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6287/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6287/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/199
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/199/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/199/comments
|
https://api.github.com/repos/ollama/ollama/issues/199/events
|
https://github.com/ollama/ollama/issues/199
| 1,819,047,422
|
I_kwDOJ0Z1Ps5sbHX-
| 199
|
nous-hermes and parameters
|
{
"login": "alivardar",
"id": 10295369,
"node_id": "MDQ6VXNlcjEwMjk1MzY5",
"avatar_url": "https://avatars.githubusercontent.com/u/10295369?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/alivardar",
"html_url": "https://github.com/alivardar",
"followers_url": "https://api.github.com/users/alivardar/followers",
"following_url": "https://api.github.com/users/alivardar/following{/other_user}",
"gists_url": "https://api.github.com/users/alivardar/gists{/gist_id}",
"starred_url": "https://api.github.com/users/alivardar/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/alivardar/subscriptions",
"organizations_url": "https://api.github.com/users/alivardar/orgs",
"repos_url": "https://api.github.com/users/alivardar/repos",
"events_url": "https://api.github.com/users/alivardar/events{/privacy}",
"received_events_url": "https://api.github.com/users/alivardar/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 4
| 2023-07-24T19:58:34
| 2023-08-23T17:46:45
| 2023-08-23T17:46:45
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello,
If I want to generate my model, this example with temperature and num_ctx paramters crashing "ollama" application.
FROM nous-hermes
# sets the temperature to 1 [higher is more creative, lower is more coherent]
# sets the context size to 4096
PARAMETER temperature 2
PARAMETER num_ctx 4096
Here is all result when app crashed.
alivardar@Alis-MacBook-Air ModelFiles % ollama serve
[GIN-debug] [WARNING] Creating an Engine instance with the Logger and Recovery middleware already attached.
[GIN-debug] [WARNING] Running in "debug" mode. Switch to "release" mode in production.
- using env: export GIN_MODE=release
- using code: gin.SetMode(gin.ReleaseMode)
[GIN-debug] GET / --> github.com/jmorganca/ollama/server.Serve.func1 (3 handlers)
[GIN-debug] POST /api/pull --> github.com/jmorganca/ollama/server.PullModelHandler (3 handlers)
[GIN-debug] POST /api/generate --> github.com/jmorganca/ollama/server.GenerateHandler (3 handlers)
[GIN-debug] POST /api/create --> github.com/jmorganca/ollama/server.CreateModelHandler (3 handlers)
[GIN-debug] POST /api/push --> github.com/jmorganca/ollama/server.PushModelHandler (3 handlers)
[GIN-debug] GET /api/tags --> github.com/jmorganca/ollama/server.ListModelsHandler (3 handlers)
[GIN-debug] DELETE /api/delete --> github.com/jmorganca/ollama/server.DeleteModelHandler (3 handlers)
2023/07/24 22:54:48 routes.go:237: Listening on 127.0.0.1:11434
llama.cpp: loading model from /Users/alivardar/.ollama/models/blobs/sha256:d1735b93e1dc503f1045ccd6c8bd73277b18ba892befd1dc29e9b9a7822ed998
llama_model_load_internal: format = ggjt v3 (latest)
llama_model_load_internal: n_vocab = 32001
llama_model_load_internal: n_ctx = 4096
llama_model_load_internal: n_embd = 5120
llama_model_load_internal: n_mult = 256
llama_model_load_internal: n_head = 40
llama_model_load_internal: n_layer = 40
llama_model_load_internal: n_rot = 128
llama_model_load_internal: freq_base = 10000.0
llama_model_load_internal: freq_scale = 1
llama_model_load_internal: ftype = 2 (mostly Q4_0)
llama_model_load_internal: n_ff = 13824
llama_model_load_internal: model size = 13B
llama_model_load_internal: ggml ctx size = 0.09 MB
llama_model_load_internal: mem required = 9132.72 MB (+ 1608.00 MB per state)
llama_new_context_with_model: kv self size = 3200.00 MB
ggml_metal_init: allocating
ggml_metal_init: using MPS
ggml_metal_init: loading '/Applications/Ollama.app/Contents/Resources/ggml-metal.metal'
ggml_metal_init: loaded kernel_add 0x152f0a280
ggml_metal_init: loaded kernel_mul 0x152f0a880
ggml_metal_init: loaded kernel_mul_row 0x152f0aeb0
ggml_metal_init: loaded kernel_scale 0x152f0b3d0
ggml_metal_init: loaded kernel_silu 0x152f0b8f0
ggml_metal_init: loaded kernel_relu 0x152f0be10
ggml_metal_init: loaded kernel_gelu 0x152f0c330
ggml_metal_init: loaded kernel_soft_max 0x152f0c9e0
ggml_metal_init: loaded kernel_diag_mask_inf 0x152f0d040
ggml_metal_init: loaded kernel_get_rows_f16 0x152f0d6c0
ggml_metal_init: loaded kernel_get_rows_q4_0 0x152f0dd40
ggml_metal_init: loaded kernel_get_rows_q4_1 0x152f0e530
ggml_metal_init: loaded kernel_get_rows_q2_K 0x152f0ebb0
ggml_metal_init: loaded kernel_get_rows_q3_K 0x152f0f230
ggml_metal_init: loaded kernel_get_rows_q4_K 0x152f0f8b0
ggml_metal_init: loaded kernel_get_rows_q5_K 0x152f0ff30
ggml_metal_init: loaded kernel_get_rows_q6_K 0x152f105b0
ggml_metal_init: loaded kernel_rms_norm 0x152f10c70
ggml_metal_init: loaded kernel_norm 0x152f11320
ggml_metal_init: loaded kernel_mul_mat_f16_f32 0x152f11cf0
ggml_metal_init: loaded kernel_mul_mat_q4_0_f32 0x152f123b0
ggml_metal_init: loaded kernel_mul_mat_q4_1_f32 0x152f12a70
ggml_metal_init: loaded kernel_mul_mat_q2_K_f32 0x152f13150
ggml_metal_init: loaded kernel_mul_mat_q3_K_f32 0x152f139d0
ggml_metal_init: loaded kernel_mul_mat_q4_K_f32 0x152f14090
ggml_metal_init: loaded kernel_mul_mat_q5_K_f32 0x152f14730
ggml_metal_init: loaded kernel_mul_mat_q6_K_f32 0x152f14dd0
ggml_metal_init: loaded kernel_rope 0x152f154f0
ggml_metal_init: loaded kernel_alibi_f32 0x152f16010
ggml_metal_init: loaded kernel_cpy_f32_f16 0x152f168a0
ggml_metal_init: loaded kernel_cpy_f32_f32 0x152f17130
ggml_metal_init: loaded kernel_cpy_f16_f16 0x152f179c0
ggml_metal_init: recommendedMaxWorkingSetSize = 10922.67 MB
ggml_metal_init: hasUnifiedMemory = true
ggml_metal_init: maxTransferRate = built-in GPU
llama_new_context_with_model: max tensor size = 87.89 MB
ggml_metal_add_buffer: allocated 'data ' buffer, size = 6984.06 MB, ( 6984.52 / 10922.67)
ggml_metal_add_buffer: allocated 'eval ' buffer, size = 1040.00 MB, ( 8024.52 / 10922.67)
ggml_metal_add_buffer: allocated 'kv ' buffer, size = 3202.00 MB, (11226.52 / 10922.67), warning: current allocated size is greater than the recommended max working set size
ggml_metal_add_buffer: allocated 'scr0 ' buffer, size = 597.00 MB, (11823.52 / 10922.67), warning: current allocated size is greater than the recommended max working set size
ggml_metal_add_buffer: allocated 'scr1 ' buffer, size = 512.00 MB, (12335.52 / 10922.67), warning: current allocated size is greater than the recommended max working set size
ggml_metal_graph_compute: command buffer 0 failed with status 5
GGML_ASSERT: ggml-metal.m:1023: false
SIGABRT: abort
PC=0x1a1500724 m=5 sigcode=0
signal arrived during cgo execution
goroutine 19 [syscall]:
runtime.cgocall(0x102d99e8c, 0x1400018d278)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/cgocall.go:157 +0x54 fp=0x1400018d240 sp=0x1400018d200 pc=0x102888c64
github.com/jmorganca/ollama/llama._Cfunc_llama_eval(0x153813c00, 0x14000471e28, 0x1, 0x0, 0x8)
_cgo_gotypes.go:210 +0x38 fp=0x1400018d270 sp=0x1400018d240 pc=0x102d87448
github.com/jmorganca/ollama/llama.New.func4(0x102fc2de0?, {0x14000471e28, 0x1, 0x0?}, {0xffffffffffffffff, 0x0, 0x1000, 0x200, 0x1, 0x0, ...})
/Users/jmorgan/workspace/ollama/llama/llama.go:141 +0x7c fp=0x1400018d2c0 sp=0x1400018d270 pc=0x102d8826c
github.com/jmorganca/ollama/llama.New({0x1400028ce00, 0x6d}, {0xffffffffffffffff, 0x0, 0x1000, 0x200, 0x1, 0x0, 0x0, 0x1, ...})
/Users/jmorgan/workspace/ollama/llama/llama.go:141 +0x288 fp=0x1400018d480 sp=0x1400018d2c0 pc=0x102d88028
github.com/jmorganca/ollama/server.GenerateHandler(0x14000498300)
/Users/jmorgan/workspace/ollama/server/routes.go:54 +0x5c0 fp=0x1400018d6e0 sp=0x1400018d480 pc=0x102d939f0
github.com/gin-gonic/gin.(*Context).Next(...)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/context.go:174
github.com/gin-gonic/gin.CustomRecoveryWithWriter.func1(0x14000498300)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/recovery.go:102 +0x7c fp=0x1400018d730 sp=0x1400018d6e0 pc=0x102d7eb3c
github.com/gin-gonic/gin.(*Context).Next(...)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/context.go:174
github.com/gin-gonic/gin.LoggerWithConfig.func1(0x14000498300)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/logger.go:240 +0xac fp=0x1400018d8e0 sp=0x1400018d730 pc=0x102d7ddbc
github.com/gin-gonic/gin.(*Context).Next(...)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/context.go:174
github.com/gin-gonic/gin.(*Engine).handleHTTPRequest(0x1400047a820, 0x14000498300)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/gin.go:620 +0x54c fp=0x1400018da70 sp=0x1400018d8e0 pc=0x102d7cecc
github.com/gin-gonic/gin.(*Engine).ServeHTTP(0x1400047a820, {0x103014f10?, 0x14000443420}, 0x14000498200)
/Users/jmorgan/go/pkg/mod/github.com/gin-gonic/gin@v1.9.1/gin.go:576 +0x1d4 fp=0x1400018dab0 sp=0x1400018da70 pc=0x102d7c7d4
net/http.serverHandler.ServeHTTP({0x103012e70?}, {0x103014f10, 0x14000443420}, 0x14000498200)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:2936 +0x2d8 fp=0x1400018db60 sp=0x1400018dab0 pc=0x102b06408
net/http.(*conn).serve(0x140001a0900, {0x103015588, 0x1400049a060})
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:1995 +0x560 fp=0x1400018dfa0 sp=0x1400018db60 pc=0x102b02100
net/http.(*Server).Serve.func3()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:3089 +0x30 fp=0x1400018dfd0 sp=0x1400018dfa0 pc=0x102b06c30
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x1400018dfd0 sp=0x1400018dfd0 pc=0x1028ec074
created by net/http.(*Server).Serve
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:3089 +0x520
goroutine 1 [IO wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:381 +0xe4 fp=0x14000337860 sp=0x14000337840 pc=0x1028bbee4
runtime.netpollblock(0x140003378f8?, 0x296fca4?, 0x1?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/netpoll.go:527 +0x158 fp=0x140003378a0 sp=0x14000337860 pc=0x1028b5408
internal/poll.runtime_pollWait(0x12aae7a18, 0x72)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/netpoll.go:306 +0xa0 fp=0x140003378d0 sp=0x140003378a0 pc=0x1028e5c40
internal/poll.(*pollDesc).wait(0x14000478600?, 0x0?, 0x0)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/internal/poll/fd_poll_runtime.go:84 +0x28 fp=0x14000337900 sp=0x140003378d0 pc=0x10296b2e8
internal/poll.(*pollDesc).waitRead(...)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/internal/poll/fd_poll_runtime.go:89
internal/poll.(*FD).Accept(0x14000478600)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/internal/poll/fd_unix.go:614 +0x250 fp=0x140003379b0 sp=0x14000337900 pc=0x10296fd90
net.(*netFD).accept(0x14000478600)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/fd_unix.go:172 +0x28 fp=0x14000337a70 sp=0x140003379b0 pc=0x1029af2f8
net.(*TCPListener).accept(0x1400012ed68)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/tcpsock_posix.go:148 +0x28 fp=0x14000337aa0 sp=0x14000337a70 pc=0x1029c48f8
net.(*TCPListener).Accept(0x1400012ed68)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/tcpsock.go:297 +0x2c fp=0x14000337ae0 sp=0x14000337aa0 pc=0x1029c3a6c
net/http.(*onceCloseListener).Accept(0x140001a0900?)
<autogenerated>:1 +0x30 fp=0x14000337b00 sp=0x14000337ae0 pc=0x102b2a3b0
net/http.(*Server).Serve(0x1400039cff0, {0x103014d00, 0x1400012ed68})
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:3059 +0x304 fp=0x14000337c30 sp=0x14000337b00 pc=0x102b068d4
github.com/jmorganca/ollama/server.Serve({0x103014d00, 0x1400012ed68})
/Users/jmorgan/workspace/ollama/server/routes.go:242 +0x294 fp=0x14000337ca0 sp=0x14000337c30 pc=0x102d95aa4
github.com/jmorganca/ollama/cmd.RunServer(0x14000449200?, {0x102deac46?, 0x0?, 0x0?})
/Users/jmorgan/workspace/ollama/cmd/cmd.go:384 +0x114 fp=0x14000337d20 sp=0x14000337ca0 pc=0x102d98ea4
github.com/spf13/cobra.(*Command).execute(0x14000449200, {0x10348e290, 0x0, 0x0})
/Users/jmorgan/go/pkg/mod/github.com/spf13/cobra@v1.7.0/command.go:940 +0x5c8 fp=0x14000337e60 sp=0x14000337d20 pc=0x102bacb18
github.com/spf13/cobra.(*Command).ExecuteC(0x14000448900)
/Users/jmorgan/go/pkg/mod/github.com/spf13/cobra@v1.7.0/command.go:1068 +0x35c fp=0x14000337f20 sp=0x14000337e60 pc=0x102bad26c
github.com/spf13/cobra.(*Command).Execute(...)
/Users/jmorgan/go/pkg/mod/github.com/spf13/cobra@v1.7.0/command.go:992
github.com/spf13/cobra.(*Command).ExecuteContext(0x14000054768?, {0x103015518?, 0x14000120010?})
/Users/jmorgan/go/pkg/mod/github.com/spf13/cobra@v1.7.0/command.go:985 +0x50 fp=0x14000337f40 sp=0x14000337f20 pc=0x102bace00
main.main()
/Users/jmorgan/workspace/ollama/main.go:10 +0x34 fp=0x14000337f70 sp=0x14000337f40 pc=0x102d99c54
runtime.main()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:250 +0x248 fp=0x14000337fd0 sp=0x14000337f70 pc=0x1028bbab8
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x14000337fd0 sp=0x14000337fd0 pc=0x1028ec074
goroutine 2 [force gc (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:381 +0xe4 fp=0x14000054fa0 sp=0x14000054f80 pc=0x1028bbee4
runtime.goparkunlock(...)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:387
runtime.forcegchelper()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:305 +0xb8 fp=0x14000054fd0 sp=0x14000054fa0 pc=0x1028bbd28
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x14000054fd0 sp=0x14000054fd0 pc=0x1028ec074
created by runtime.init.6
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:293 +0x24
goroutine 3 [GC sweep wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:381 +0xe4 fp=0x14000055760 sp=0x14000055740 pc=0x1028bbee4
runtime.goparkunlock(...)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:387
runtime.bgsweep(0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgcsweep.go:278 +0xa4 fp=0x140000557b0 sp=0x14000055760 pc=0x1028a8bc4
runtime.gcenable.func1()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgc.go:178 +0x28 fp=0x140000557d0 sp=0x140000557b0 pc=0x10289d6d8
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x140000557d0 sp=0x140000557d0 pc=0x1028ec074
created by runtime.gcenable
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgc.go:178 +0x74
goroutine 4 [GC scavenge wait]:
runtime.gopark(0x1400007c000?, 0x102ee86e8?, 0x1?, 0x0?, 0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:381 +0xe4 fp=0x14000055f50 sp=0x14000055f30 pc=0x1028bbee4
runtime.goparkunlock(...)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:387
runtime.(*scavengerState).park(0x1033d2d60)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgcscavenge.go:400 +0x5c fp=0x14000055f80 sp=0x14000055f50 pc=0x1028a6a3c
runtime.bgscavenge(0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgcscavenge.go:628 +0x44 fp=0x14000055fb0 sp=0x14000055f80 pc=0x1028a6fb4
runtime.gcenable.func2()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgc.go:179 +0x28 fp=0x14000055fd0 sp=0x14000055fb0 pc=0x10289d678
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x14000055fd0 sp=0x14000055fd0 pc=0x1028ec074
created by runtime.gcenable
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mgc.go:179 +0xb8
goroutine 18 [finalizer wait]:
runtime.gopark(0x1a0?, 0x1033d37a0?, 0x40?, 0x23?, 0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:381 +0xe4 fp=0x14000054580 sp=0x14000054560 pc=0x1028bbee4
runtime.runfinq()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mfinal.go:193 +0x10c fp=0x140000547d0 sp=0x14000054580 pc=0x10289c76c
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x140000547d0 sp=0x140000547d0 pc=0x1028ec074
created by runtime.createfing
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/mfinal.go:163 +0x84
goroutine 20 [IO wait]:
runtime.gopark(0xffffffffffffffff?, 0xffffffffffffffff?, 0x23?, 0x0?, 0x1028fe890?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/proc.go:381 +0xe4 fp=0x14000050540 sp=0x14000050520 pc=0x1028bbee4
runtime.netpollblock(0x0?, 0x0?, 0x0?)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/netpoll.go:527 +0x158 fp=0x14000050580 sp=0x14000050540 pc=0x1028b5408
internal/poll.runtime_pollWait(0x12aae7928, 0x72)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/netpoll.go:306 +0xa0 fp=0x140000505b0 sp=0x14000050580 pc=0x1028e5c40
internal/poll.(*pollDesc).wait(0x14000478680?, 0x1400049a161?, 0x0)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/internal/poll/fd_poll_runtime.go:84 +0x28 fp=0x140000505e0 sp=0x140000505b0 pc=0x10296b2e8
internal/poll.(*pollDesc).waitRead(...)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/internal/poll/fd_poll_runtime.go:89
internal/poll.(*FD).Read(0x14000478680, {0x1400049a161, 0x1, 0x1})
/opt/homebrew/Cellar/go/1.20.5/libexec/src/internal/poll/fd_unix.go:167 +0x200 fp=0x14000050680 sp=0x140000505e0 pc=0x10296c650
net.(*netFD).Read(0x14000478680, {0x1400049a161?, 0x0?, 0x0?})
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/fd_posix.go:55 +0x28 fp=0x140000506d0 sp=0x14000050680 pc=0x1029ad658
net.(*conn).Read(0x14000130d00, {0x1400049a161?, 0x0?, 0x0?})
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/net.go:183 +0x34 fp=0x14000050720 sp=0x140000506d0 pc=0x1029bbc64
net.(*TCPConn).Read(0x0?, {0x1400049a161?, 0x0?, 0x0?})
<autogenerated>:1 +0x2c fp=0x14000050750 sp=0x14000050720 pc=0x1029ce1ac
net/http.(*connReader).backgroundRead(0x1400049a150)
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:674 +0x44 fp=0x140000507b0 sp=0x14000050750 pc=0x102afc5b4
net/http.(*connReader).startBackgroundRead.func2()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:670 +0x28 fp=0x140000507d0 sp=0x140000507b0 pc=0x102afc4d8
runtime.goexit()
/opt/homebrew/Cellar/go/1.20.5/libexec/src/runtime/asm_arm64.s:1172 +0x4 fp=0x140000507d0 sp=0x140000507d0 pc=0x1028ec074
created by net/http.(*connReader).startBackgroundRead
/opt/homebrew/Cellar/go/1.20.5/libexec/src/net/http/server.go:670 +0xcc
r0 0x0
r1 0x0
r2 0x0
r3 0x0
r4 0x0
r5 0x16f586c00
r6 0xa
r7 0x0
r8 0xe035c4e8feb62ded
r9 0xe035c4e991efdded
r10 0x2
r11 0xfffffffd
r12 0x10000000000
r13 0x0
r14 0x0
r15 0x0
r16 0x148
r17 0x201034f60
r18 0x0
r19 0x6
r20 0x16f59f000
r21 0x1603
r22 0x16f59f0e0
r23 0x8
r24 0x7
r25 0x8
r26 0x1fc513720
r27 0x102ddb9c0
r28 0x102f010f0
r29 0x16f586bb0
lr 0x1a1537c28
sp 0x16f586b90
pc 0x1a1500724
fault 0x1a1500724
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/199/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/199/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8657
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8657/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8657/comments
|
https://api.github.com/repos/ollama/ollama/issues/8657/events
|
https://github.com/ollama/ollama/issues/8657
| 2,818,103,966
|
I_kwDOJ0Z1Ps6n-Nqe
| 8,657
|
running ollama deepseek-r1:1.5b on windows stuck for whole day
|
{
"login": "aadltya",
"id": 142524039,
"node_id": "U_kgDOCH6-hw",
"avatar_url": "https://avatars.githubusercontent.com/u/142524039?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/aadltya",
"html_url": "https://github.com/aadltya",
"followers_url": "https://api.github.com/users/aadltya/followers",
"following_url": "https://api.github.com/users/aadltya/following{/other_user}",
"gists_url": "https://api.github.com/users/aadltya/gists{/gist_id}",
"starred_url": "https://api.github.com/users/aadltya/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/aadltya/subscriptions",
"organizations_url": "https://api.github.com/users/aadltya/orgs",
"repos_url": "https://api.github.com/users/aadltya/repos",
"events_url": "https://api.github.com/users/aadltya/events{/privacy}",
"received_events_url": "https://api.github.com/users/aadltya/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677370291,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCVsw",
"url": "https://api.github.com/repos/ollama/ollama/labels/networking",
"name": "networking",
"color": "0B5368",
"default": false,
"description": "Issues relating to ollama pull and push"
}
] |
closed
| false
| null |
[] | null | 3
| 2025-01-29T12:40:29
| 2025-01-29T13:44:51
| 2025-01-29T13:44:50
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
optimize for low end device, I'm using should windows with 8gb ram and 4gb nvidia gtx 1650 graphics card and im unable to run PS deepseek-r1:1.5b
In command line it stuck at 0% for whole day
```bash
C:\Users\ADITYA> ollama run deepseek-r1:1.5b
pulling manifest
pulling aabd4debf0c8... 0% ▕ ▏ 0 B/1.1 GB
```

|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8657/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8657/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8267
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8267/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8267/comments
|
https://api.github.com/repos/ollama/ollama/issues/8267/events
|
https://github.com/ollama/ollama/pull/8267
| 2,762,422,303
|
PR_kwDOJ0Z1Ps6GZHcU
| 8,267
|
examples: remove codified examples
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-12-29T22:10:00
| 2025-01-13T19:26:25
| 2025-01-13T19:26:23
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/8267",
"html_url": "https://github.com/ollama/ollama/pull/8267",
"diff_url": "https://github.com/ollama/ollama/pull/8267.diff",
"patch_url": "https://github.com/ollama/ollama/pull/8267.patch",
"merged_at": "2025-01-13T19:26:22"
}
|
This PR aims to streamline the examples and to have outgoing links to community frameworks instead.
Closes #8117
|
{
"login": "ParthSareen",
"id": 29360864,
"node_id": "MDQ6VXNlcjI5MzYwODY0",
"avatar_url": "https://avatars.githubusercontent.com/u/29360864?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParthSareen",
"html_url": "https://github.com/ParthSareen",
"followers_url": "https://api.github.com/users/ParthSareen/followers",
"following_url": "https://api.github.com/users/ParthSareen/following{/other_user}",
"gists_url": "https://api.github.com/users/ParthSareen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParthSareen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParthSareen/subscriptions",
"organizations_url": "https://api.github.com/users/ParthSareen/orgs",
"repos_url": "https://api.github.com/users/ParthSareen/repos",
"events_url": "https://api.github.com/users/ParthSareen/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParthSareen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8267/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8267/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5090
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5090/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5090/comments
|
https://api.github.com/repos/ollama/ollama/issues/5090/events
|
https://github.com/ollama/ollama/issues/5090
| 2,356,059,419
|
I_kwDOJ0Z1Ps6Mbp0b
| 5,090
|
`amdgpu version file missing` when running via systemd
|
{
"login": "pulpocaminante",
"id": 109849915,
"node_id": "U_kgDOBowtOw",
"avatar_url": "https://avatars.githubusercontent.com/u/109849915?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pulpocaminante",
"html_url": "https://github.com/pulpocaminante",
"followers_url": "https://api.github.com/users/pulpocaminante/followers",
"following_url": "https://api.github.com/users/pulpocaminante/following{/other_user}",
"gists_url": "https://api.github.com/users/pulpocaminante/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pulpocaminante/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pulpocaminante/subscriptions",
"organizations_url": "https://api.github.com/users/pulpocaminante/orgs",
"repos_url": "https://api.github.com/users/pulpocaminante/repos",
"events_url": "https://api.github.com/users/pulpocaminante/events{/privacy}",
"received_events_url": "https://api.github.com/users/pulpocaminante/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6433346500,
"node_id": "LA_kwDOJ0Z1Ps8AAAABf3UTxA",
"url": "https://api.github.com/repos/ollama/ollama/labels/amd",
"name": "amd",
"color": "000000",
"default": false,
"description": "Issues relating to AMD GPUs and ROCm"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-06-16T23:56:40
| 2024-06-18T19:01:33
| 2024-06-18T19:01:33
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Previous issue was closed as fixed but the bug still exists.
Hi, this doesn't happen to me when running ollama as root directly in a shell, but it happens when I start ollama as a service (regardless of the user):
```
amnesia λ ~/ sudo systemctl status ollama
● ollama.service - Ollama Service
Loaded: loaded (/etc/systemd/system/ollama.service; enabled; preset: disabled)
Active: active (running) since Sun 2024-06-16 16:47:04 PDT; 39s ago
Main PID: 7273 (ollama)
Tasks: 18 (limit: 38365)
Memory: 561.0M (peak: 615.8M)
CPU: 5.644s
CGroup: /system.slice/ollama.service
└─7273 /usr/local/bin/ollama serve
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.252-07:00 level=INFO source=amd_linux.go:304 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION="\"10.3.0\""
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.506-07:00 level=WARN source=amd_linux.go:48 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sy>
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.511-07:00 level=INFO source=amd_linux.go:304 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION="\"10.3.0\""
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.714-07:00 level=WARN source=sched.go:511 msg="gpu VRAM usage didn't recover within timeout" seconds=5.045401852
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.749-07:00 level=WARN source=amd_linux.go:48 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sy>
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.753-07:00 level=INFO source=amd_linux.go:304 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION="\"10.3.0\""
Jun 16 16:47:37 dead ollama[7273]: time=2024-06-16T16:47:37.964-07:00 level=WARN source=sched.go:511 msg="gpu VRAM usage didn't recover within timeout" seconds=5.295260429
Jun 16 16:47:38 dead ollama[7273]: time=2024-06-16T16:47:38.007-07:00 level=WARN source=amd_linux.go:48 msg="ollama recommends running the https://www.amd.com/en/support/linux-drivers" error="amdgpu version file missing: /sy>
Jun 16 16:47:38 dead ollama[7273]: time=2024-06-16T16:47:38.012-07:00 level=INFO source=amd_linux.go:304 msg="skipping rocm gfx compatibility check" HSA_OVERRIDE_GFX_VERSION="\"10.3.0\""
Jun 16 16:47:38 dead ollama[7273]: time=2024-06-16T16:47:38.214-07:00 level=WARN source=sched.go:511 msg="gpu VRAM usage didn't recover within timeout" seconds=5.545797232
```
But somehow:
`amnesia λ ~/ sudo ROCR_VISIBLE_DEVICES=0 HSA_OVERRIDE_GFX_VERSION="10.3.0" OLLAMA_DEBUG=1 ollama serve`
Works fine and I can chat without issue. Here's my service file, please note I have tried with both the ollama user and the root user (and the ollama user is properly configured/in render & video group):
```
[Unit]
Description=Ollama Service
After=network-online.target
[Service]
ExecStart=/usr/local/bin/ollama serve
User=root
Group=root
Restart=always
RestartSec=3
Environment="PATH=/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/lib/jvm/default/bin:/usr/bin/site_perl:/usr/bin/vendor_perl:/usr/bin/core_perl:/opt/rocm/bin:/usr/local/lib/baresip/modules"
Environment="ROCR_VISIBLE_DEVICES=0"
Environment="HSA_OVERRIDE_GFX_VERSION=\"10.3.0\""
[Install]
WantedBy=default.target
```
Both in the shell & run as a service they report using the same GPU (id=0, 6700XT):
`level=INFO source=amd_linux.go:71 msg="inference compute" id=0 library=rocm compute=gfx1031 driver=0.0 name=1002:73df total="12.0 GiB" available="12.0 GiB"`
_Originally posted by @pulpocaminante in https://github.com/ollama/ollama/issues/4427#issuecomment-2171941948_
|
{
"login": "pulpocaminante",
"id": 109849915,
"node_id": "U_kgDOBowtOw",
"avatar_url": "https://avatars.githubusercontent.com/u/109849915?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pulpocaminante",
"html_url": "https://github.com/pulpocaminante",
"followers_url": "https://api.github.com/users/pulpocaminante/followers",
"following_url": "https://api.github.com/users/pulpocaminante/following{/other_user}",
"gists_url": "https://api.github.com/users/pulpocaminante/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pulpocaminante/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pulpocaminante/subscriptions",
"organizations_url": "https://api.github.com/users/pulpocaminante/orgs",
"repos_url": "https://api.github.com/users/pulpocaminante/repos",
"events_url": "https://api.github.com/users/pulpocaminante/events{/privacy}",
"received_events_url": "https://api.github.com/users/pulpocaminante/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5090/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5090/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/684
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/684/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/684/comments
|
https://api.github.com/repos/ollama/ollama/issues/684/events
|
https://github.com/ollama/ollama/issues/684
| 1,923,008,644
|
I_kwDOJ0Z1Ps5ynsiE
| 684
|
WSL2 Ubuntu 22.04 GPU "CUDA error 100" ggml-cuda.cu:5522 ggml-cuda.cu:4883 no CUDA-capable device is detected
|
{
"login": "iamexe",
"id": 60526252,
"node_id": "MDQ6VXNlcjYwNTI2MjUy",
"avatar_url": "https://avatars.githubusercontent.com/u/60526252?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iamexe",
"html_url": "https://github.com/iamexe",
"followers_url": "https://api.github.com/users/iamexe/followers",
"following_url": "https://api.github.com/users/iamexe/following{/other_user}",
"gists_url": "https://api.github.com/users/iamexe/gists{/gist_id}",
"starred_url": "https://api.github.com/users/iamexe/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/iamexe/subscriptions",
"organizations_url": "https://api.github.com/users/iamexe/orgs",
"repos_url": "https://api.github.com/users/iamexe/repos",
"events_url": "https://api.github.com/users/iamexe/events{/privacy}",
"received_events_url": "https://api.github.com/users/iamexe/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 14
| 2023-10-02T23:45:31
| 2024-01-21T09:53:28
| 2023-10-03T23:44:47
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Thank you so much for ollama and the wsl2 support,
I already wrote a vuejs frontend and it works great with CPU.
I want GPU on WSL.
I installed CUDA like recomended from nvidia with wsl2 (cuda on windows).
I ran the following:
go generate ./...
go build .
I got a ollama that runs with CPU but not with GPU.
In journalctl | grep cuda I see:
/home/y/Dev/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:5522: no CUDA-capable device is detected
Everytime I run any model in ollama I get that error. I tried with mistral and with my own gpu-mistral that had num_gpu 50 and the same with num_gpu 1000. It doesn't matter. I am able to create the models with num_gpu.
When I run ollama/llm/llama.cpp/gguf/build/cuda/bin/server or ggml/...../server it used to have the same error I am still facing with ollama now.
I adjusted my environmennt variables and now the error doesn't show anymore for those built "server" binaries. ollama still shows the error.
I adjusted my enviroment variables like this:
**cat /etc/*/environment_variables.sh**
```
export CUDA_PATH="/usr/local/cuda-12.2/bin"
export LD_LIBRARY_PATH="/mnt/c/Windows/System32/lxss/lib:/usr/local/cuda/lib64:/usr/local/cuda/lib64/stubs:/usr/lib/x86_64-linux-gnu"
export PATH=/usr/local/cuda-12.2/bin${PATH:+:${PATH}}
```
It works fine for the server binaries of gguf and ggml (the error 100 was no longer present there).
Info: There is no more text from ehre on apart of console commands and their output.
**./Dev/ollama/llm/llama.cpp/gguf/build/cuda/bin/server**
```
ggml_init_cublas: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 2060 with Max-Q Design, compute capability 7.5
{"timestamp":1696288901,"level":"INFO","function":"main","line":1294,"message":"build info","build":1267,"commit":"bc9d3e3"}
{"timestamp":1696288901,"level":"INFO","function":"main","line":1296,"message":"system info","n_threads":8,"total_threads":16,"system_info":"AVX = 1 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | "}
error loading model: failed to open models/7B/ggml-model-f16.gguf: No such file or directory
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model 'models/7B/ggml-model-f16.gguf'
{"timestamp":1696288901,"level":"ERROR","function":"loadModel","line":265,"message":"unable to load model","model":"models/7B/ggml-model-f16.gguf"}
```
**./Dev/ollama/llm/llama.cpp/ggml/build/cuda/bin/server**
```
ggml_init_cublas: found 1 CUDA devices:
Device 0: NVIDIA GeForce RTX 2060 with Max-Q Design, compute capability 7.5
{"timestamp":1696288885,"level":"INFO","function":"main","line":1190,"message":"build info","build":1009,"commit":"9e232f0"}
{"timestamp":1696288885,"level":"INFO","function":"main","line":1192,"message":"system info","n_threads":8,"total_threads":16,"system_info":"AVX = 1 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | VSX = 0 | "}
error loading model: failed to open models/7B/ggml-model.bin: No such file or directory
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model 'models/7B/ggml-model.bin'
{"timestamp":1696288885,"level":"ERROR","function":"loadModel","line":261,"message":"unable to load model","model":"models/7B/ggml-model.bin"}
```
**cmake --version**
cmake version 3.27.6
**go version**
go version go1.21.1 linux/amd64
**gcc --version**
gcc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0
**echo $PATH** (full env output with full path output at bottom)
/usr/local/cuda-12.2/bin
/usr/lib/wsl/lib
/mnt/c/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.2/bin
/mnt/c/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.2/libnvvp
/mnt/c/program files/python311/scripts/
/mnt/c/program files/python311/
/mnt/c/program files/nvidia corporation/nvidia nvdlisr
/mnt/c/Program Files (x86)/NVIDIA Corporation/PhysX/Common
/mnt/c/Program Files/NVIDIA Corporation/Nsight Compute 2023.2.2/
**Other relevant environment variables:** (full env at bottom)
LD_LIBRARY_PATH=/mnt/c/Windows/System32/lxss/lib:/usr/local/cuda/lib64:/usr/local/cuda/lib64/stubs:/usr/lib/x86_64-linux-gnu
WSL2_GUI_APPS_ENABLED=1
WSL_DISTRO_NAME=Ubuntu-22.04
CUDA_PATH=/usr/local/cuda-12.2/bin
**journalctl | grep cuda**
```
Oct 03 01:10:56 c1 unknown: /usr/lib/wsl/lib/libcuda.so.1 is not a symbolic link
Oct 03 01:12:30 c1 ollama[884]: CUDA error 100 at /home/y/Dev/ollama/llm/llama.cpp/gguf/ggml-cuda.cu:5522: no CUDA-capable device is detected
Oct 03 01:15:33 c1 ollama[1138]: CUDA error 100 at /home/y/Dev/ollama/llm/llama.cpp/ggml/ggml-cuda.cu:4883: no CUDA-capable device is detected
```
**nvcc --version**
```
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Tue_Aug_15_22:02:13_PDT_2023
Cuda compilation tools, release 12.2, V12.2.140
Build cuda_12.2.r12.2/compiler.33191640_0
```
**nvidia-smi**
Tue Oct 3 01:18:06 2023
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 535.112 Driver Version: 537.42 CUDA Version: 12.2 |
|-----------------------------------------+----------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 NVIDIA GeForce RTX 2060 ... On | 00000000:01:00.0 Off | N/A |
| N/A 54C P8 4W / 65W | 12MiB / 6144MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
+---------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=======================================================================================|
| No running processes found |
+---------------------------------------------------------------------------------------+
**env**
```
SHELL=/bin/bash
NVM_INC=/home/y/.nvm/versions/node/v20.6.1/include/node
WSL2_GUI_APPS_ENABLED=1
WSL_DISTRO_NAME=Ubuntu-22.04
NAME=c1
PWD=/home/y
LOGNAME=y
HOME=/home/y
LANG=C.UTF-8
WSL_INTEROP=/run/WSL/391_interop
LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=30;41:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:
WAYLAND_DISPLAY=wayland-0
NVM_DIR=/home/y/.nvm
LESSCLOSE=/usr/bin/lesspipe %s %s
TERM=xterm-256color
LESSOPEN=| /usr/bin/lesspipe %s
USER=y
CUDA_PATH=/usr/local/cuda-12.2/bin
DISPLAY=:0
SHLVL=1
NVM_CD_FLAGS=
LD_LIBRARY_PATH=/mnt/c/Windows/System32/lxss/lib:/usr/local/cuda/lib64:/usr/local/cuda/lib64/stubs:/usr/lib/x86_64-linux-gnu
XDG_RUNTIME_DIR=/run/user/1000/
WSLENV=
XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
PATH=/home/y/.local/bin:/home/y/.nvm/versions/node/v20.6.1/bin:/usr/local/cuda-12.2/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/usr/lib/wsl/lib:/mnt/c/Program Files/WindowsApps/CanonicalGroupLimited.Ubuntu22.04LTS_2204.2.47.0_x64__79rhkp1fndgsc:/mnt/c/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.2/bin:/mnt/c/Program Files/NVIDIA GPU Computing Toolkit/CUDA/v12.2/libnvvp:/mnt/c/program files/python311/scripts/:/mnt/c/program files/python311/:/mnt/c/program files/common files/oracle/java/javapath:/mnt/c/windows/system32:/mnt/c/windows:/mnt/c/windows/system32/wbem:/mnt/c/windows/system32/windowspowershell/v1.0/:/mnt/c/windows/system32/openssh/:/mnt/c/program files/dotnet/:/mnt/c/programdata/chocolatey/bin:/mnt/c/program files/microsoft vs code/bin:/mnt/c/program files/putty/:/mnt/c/program files/nvidia corporation/nvidia nvdlisr:/mnt/c/program files (x86)/vim/vim82/:/mnt/c/windows/system32/openssh/:/mnt/c/program files/nodejs/:/mnt/c/program files/process lasso/:/mnt/c/Program Files/PowerShell/7/:/mnt/c/Program Files (x86)/NVIDIA Corporation/PhysX/Common:/mnt/c/Program Files/NVIDIA Corporation/Nsight Compute 2023.2.2/:/mnt/c/Users/User/AppData/Local/Microsoft/WindowsApps:/mnt/c/Windows/Microsoft.NET/Framework/v4.0.30319/:/mnt/c/Program Files (x86)/Vim/vim82/vim.exe:/mnt/c/Program Files/Java/jdk-19/bin/java.exe:/mnt/c/Users/User/AppData/Roaming/npm:/mnt/c/Users/User/AppData/Local/GitHubDesktop/bin:/mnt/c/Program Files (x86)/Nmap:/snap/bin
DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1000/bus
NVM_BIN=/home/y/.nvm/versions/node/v20.6.1/bin
HOSTTYPE=x86_64
PULSE_SERVER=unix:/mnt/wslg/PulseServer
_=/usr/bin/env
```
So am I missing something? Thank you for any hints!
|
{
"login": "iamexe",
"id": 60526252,
"node_id": "MDQ6VXNlcjYwNTI2MjUy",
"avatar_url": "https://avatars.githubusercontent.com/u/60526252?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iamexe",
"html_url": "https://github.com/iamexe",
"followers_url": "https://api.github.com/users/iamexe/followers",
"following_url": "https://api.github.com/users/iamexe/following{/other_user}",
"gists_url": "https://api.github.com/users/iamexe/gists{/gist_id}",
"starred_url": "https://api.github.com/users/iamexe/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/iamexe/subscriptions",
"organizations_url": "https://api.github.com/users/iamexe/orgs",
"repos_url": "https://api.github.com/users/iamexe/repos",
"events_url": "https://api.github.com/users/iamexe/events{/privacy}",
"received_events_url": "https://api.github.com/users/iamexe/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/684/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/684/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2903
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2903/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2903/comments
|
https://api.github.com/repos/ollama/ollama/issues/2903/events
|
https://github.com/ollama/ollama/issues/2903
| 2,165,553,361
|
I_kwDOJ0Z1Ps6BE7jR
| 2,903
|
msg="CPU does not have AVX or AVX2, disabling GPU support."
|
{
"login": "digicr",
"id": 162058985,
"node_id": "U_kgDOCajS6Q",
"avatar_url": "https://avatars.githubusercontent.com/u/162058985?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/digicr",
"html_url": "https://github.com/digicr",
"followers_url": "https://api.github.com/users/digicr/followers",
"following_url": "https://api.github.com/users/digicr/following{/other_user}",
"gists_url": "https://api.github.com/users/digicr/gists{/gist_id}",
"starred_url": "https://api.github.com/users/digicr/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/digicr/subscriptions",
"organizations_url": "https://api.github.com/users/digicr/orgs",
"repos_url": "https://api.github.com/users/digicr/repos",
"events_url": "https://api.github.com/users/digicr/events{/privacy}",
"received_events_url": "https://api.github.com/users/digicr/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 4
| 2024-03-03T20:57:38
| 2024-03-06T16:49:29
| 2024-03-06T16:49:28
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
winserver2022 old cpuX5675 GPU RTX3070 CUDA11.8
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2903/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2903/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/8636
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8636/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8636/comments
|
https://api.github.com/repos/ollama/ollama/issues/8636/events
|
https://github.com/ollama/ollama/issues/8636
| 2,815,799,891
|
I_kwDOJ0Z1Ps6n1bJT
| 8,636
|
Upload compressed package file, unable to decompress and error reported
|
{
"login": "terling",
"id": 174825001,
"node_id": "U_kgDOCmueKQ",
"avatar_url": "https://avatars.githubusercontent.com/u/174825001?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/terling",
"html_url": "https://github.com/terling",
"followers_url": "https://api.github.com/users/terling/followers",
"following_url": "https://api.github.com/users/terling/following{/other_user}",
"gists_url": "https://api.github.com/users/terling/gists{/gist_id}",
"starred_url": "https://api.github.com/users/terling/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/terling/subscriptions",
"organizations_url": "https://api.github.com/users/terling/orgs",
"repos_url": "https://api.github.com/users/terling/repos",
"events_url": "https://api.github.com/users/terling/events{/privacy}",
"received_events_url": "https://api.github.com/users/terling/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 2
| 2025-01-28T14:13:01
| 2025-01-29T23:29:46
| 2025-01-29T23:29:45
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Thanks for this great program, I love it! However, I uploaded a compressed package containing the project source code in the dialog interface, and an error occurred when the program was run. Can this problem be solved?

### OS
_No response_
### GPU
_No response_
### CPU
_No response_
### Ollama version
_No response_
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8636/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8636/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3477
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3477/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3477/comments
|
https://api.github.com/repos/ollama/ollama/issues/3477/events
|
https://github.com/ollama/ollama/issues/3477
| 2,223,395,529
|
I_kwDOJ0Z1Ps6EhlLJ
| 3,477
|
Support CLIP in LLaVA to provide services externally
|
{
"login": "Andiedie",
"id": 21376471,
"node_id": "MDQ6VXNlcjIxMzc2NDcx",
"avatar_url": "https://avatars.githubusercontent.com/u/21376471?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Andiedie",
"html_url": "https://github.com/Andiedie",
"followers_url": "https://api.github.com/users/Andiedie/followers",
"following_url": "https://api.github.com/users/Andiedie/following{/other_user}",
"gists_url": "https://api.github.com/users/Andiedie/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Andiedie/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Andiedie/subscriptions",
"organizations_url": "https://api.github.com/users/Andiedie/orgs",
"repos_url": "https://api.github.com/users/Andiedie/repos",
"events_url": "https://api.github.com/users/Andiedie/events{/privacy}",
"received_events_url": "https://api.github.com/users/Andiedie/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
open
| false
| null |
[] | null | 1
| 2024-04-03T16:35:03
| 2024-04-19T15:41:21
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What are you trying to do?
I am new to ollama (including llama.cpp, of course), so my questions may be a bit silly.
My use case is to serve both CLIP and LLaVA (which combines clip and mistral) at the same time.
LLaVA can run perfectly on ollama. But I need to open another service for CLIP.
What I want to ask is
1. Can ollama support the CLIP embedding interface? The current embedding interface seems to only support text.
2. Since the frozen CLIP included in LLaVA running on ollama, can it be directly reused instead of loading two copies in memory?
### How should we solve this?
_No response_
### What is the impact of not solving this?
_No response_
### Anything else?
_No response_
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3477/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3477/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/8629
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8629/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8629/comments
|
https://api.github.com/repos/ollama/ollama/issues/8629/events
|
https://github.com/ollama/ollama/issues/8629
| 2,815,526,057
|
I_kwDOJ0Z1Ps6n0YSp
| 8,629
|
Choose path to install on Windows
|
{
"login": "EvgeniGenchev",
"id": 59848681,
"node_id": "MDQ6VXNlcjU5ODQ4Njgx",
"avatar_url": "https://avatars.githubusercontent.com/u/59848681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/EvgeniGenchev",
"html_url": "https://github.com/EvgeniGenchev",
"followers_url": "https://api.github.com/users/EvgeniGenchev/followers",
"following_url": "https://api.github.com/users/EvgeniGenchev/following{/other_user}",
"gists_url": "https://api.github.com/users/EvgeniGenchev/gists{/gist_id}",
"starred_url": "https://api.github.com/users/EvgeniGenchev/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/EvgeniGenchev/subscriptions",
"organizations_url": "https://api.github.com/users/EvgeniGenchev/orgs",
"repos_url": "https://api.github.com/users/EvgeniGenchev/repos",
"events_url": "https://api.github.com/users/EvgeniGenchev/events{/privacy}",
"received_events_url": "https://api.github.com/users/EvgeniGenchev/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 1
| 2025-01-28T12:31:56
| 2025-01-28T21:31:28
| 2025-01-28T21:31:28
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
The title is pretty self-expanatory. I would be nice to chose a folder where ollama is being installed on windows instead of defaulting to C:\Users\...
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8629/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8629/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5995
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5995/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5995/comments
|
https://api.github.com/repos/ollama/ollama/issues/5995/events
|
https://github.com/ollama/ollama/pull/5995
| 2,432,956,077
|
PR_kwDOJ0Z1Ps52nnLa
| 5,995
|
return tool calls finish reason for openai
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-07-26T20:46:29
| 2024-07-30T08:51:13
| 2024-07-29T20:56:57
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5995",
"html_url": "https://github.com/ollama/ollama/pull/5995",
"diff_url": "https://github.com/ollama/ollama/pull/5995.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5995.patch",
"merged_at": "2024-07-29T20:56:57"
}
| null |
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5995/reactions",
"total_count": 4,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 2,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5995/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/5982
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5982/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5982/comments
|
https://api.github.com/repos/ollama/ollama/issues/5982/events
|
https://github.com/ollama/ollama/issues/5982
| 2,432,159,409
|
I_kwDOJ0Z1Ps6Q986x
| 5,982
|
Ollama is amazing!!
|
{
"login": "robertguss",
"id": 5605310,
"node_id": "MDQ6VXNlcjU2MDUzMTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/5605310?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/robertguss",
"html_url": "https://github.com/robertguss",
"followers_url": "https://api.github.com/users/robertguss/followers",
"following_url": "https://api.github.com/users/robertguss/following{/other_user}",
"gists_url": "https://api.github.com/users/robertguss/gists{/gist_id}",
"starred_url": "https://api.github.com/users/robertguss/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/robertguss/subscriptions",
"organizations_url": "https://api.github.com/users/robertguss/orgs",
"repos_url": "https://api.github.com/users/robertguss/repos",
"events_url": "https://api.github.com/users/robertguss/events{/privacy}",
"received_events_url": "https://api.github.com/users/robertguss/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 6
| 2024-07-26T12:41:49
| 2024-08-25T18:42:26
| 2024-08-25T18:42:26
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
This really isn't an issue but I just wanted to say that everyone who works on and maintains this project is doing incredible work! Thank you so much for all of the countless hours and hard work you put into making Ollama.
I was a little shocked to see the project has over 900 issues at the time of this writing and I just want to hopefully encourage you all. You truly are doing important work and this project means a lot to a lot of people!
God speed.
Cheers,
Robert
|
{
"login": "robertguss",
"id": 5605310,
"node_id": "MDQ6VXNlcjU2MDUzMTA=",
"avatar_url": "https://avatars.githubusercontent.com/u/5605310?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/robertguss",
"html_url": "https://github.com/robertguss",
"followers_url": "https://api.github.com/users/robertguss/followers",
"following_url": "https://api.github.com/users/robertguss/following{/other_user}",
"gists_url": "https://api.github.com/users/robertguss/gists{/gist_id}",
"starred_url": "https://api.github.com/users/robertguss/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/robertguss/subscriptions",
"organizations_url": "https://api.github.com/users/robertguss/orgs",
"repos_url": "https://api.github.com/users/robertguss/repos",
"events_url": "https://api.github.com/users/robertguss/events{/privacy}",
"received_events_url": "https://api.github.com/users/robertguss/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5982/reactions",
"total_count": 10,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 10,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5982/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1818
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1818/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1818/comments
|
https://api.github.com/repos/ollama/ollama/issues/1818/events
|
https://github.com/ollama/ollama/pull/1818
| 2,068,204,377
|
PR_kwDOJ0Z1Ps5jW8ln
| 1,818
|
fix(cmd): history in alt prompt
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-01-05T23:58:04
| 2024-01-08T21:48:35
| 2024-01-08T21:48:35
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1818",
"html_url": "https://github.com/ollama/ollama/pull/1818",
"diff_url": "https://github.com/ollama/ollama/pull/1818.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1818.patch",
"merged_at": "2024-01-08T21:48:35"
}
|
using up/down arrows (for history) messes up multiline string inputs by replacing the alt prefix `...` with the default prefix `>>>`
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1818/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1818/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3794
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3794/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3794/comments
|
https://api.github.com/repos/ollama/ollama/issues/3794/events
|
https://github.com/ollama/ollama/issues/3794
| 2,254,988,005
|
I_kwDOJ0Z1Ps6GaGLl
| 3,794
|
模型下载最后1%速度骤降,导致下载时间超长。The download speed suddenly drops at the last 1%, resulting in an extremely long download time.
|
{
"login": "aohanhongzhi",
"id": 37319319,
"node_id": "MDQ6VXNlcjM3MzE5MzE5",
"avatar_url": "https://avatars.githubusercontent.com/u/37319319?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/aohanhongzhi",
"html_url": "https://github.com/aohanhongzhi",
"followers_url": "https://api.github.com/users/aohanhongzhi/followers",
"following_url": "https://api.github.com/users/aohanhongzhi/following{/other_user}",
"gists_url": "https://api.github.com/users/aohanhongzhi/gists{/gist_id}",
"starred_url": "https://api.github.com/users/aohanhongzhi/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/aohanhongzhi/subscriptions",
"organizations_url": "https://api.github.com/users/aohanhongzhi/orgs",
"repos_url": "https://api.github.com/users/aohanhongzhi/repos",
"events_url": "https://api.github.com/users/aohanhongzhi/events{/privacy}",
"received_events_url": "https://api.github.com/users/aohanhongzhi/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 15
| 2024-04-21T09:34:13
| 2025-01-23T23:04:23
| 2024-04-30T19:20:29
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
模型无论大小,每次下载前面99%都最大的速度可以达到29MB/s。但是最后1%速度就只有几百 KB/s。很奇怪。是进度条有问题,还是啥bug?这在我本地电脑和线上服务器都出现了。
Regardless of model size, in most cases, the download speed reaches 29MB/s for about 99% of the time before completion. However, the last 1% takes only a few hundred KB/s. This is quite strange. Is it an issue with the progress bar or some bug? This phenomenon has been observed on both my local computer and online server.



### OS
Linux
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.32
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3794/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3794/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2468
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2468/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2468/comments
|
https://api.github.com/repos/ollama/ollama/issues/2468/events
|
https://github.com/ollama/ollama/pull/2468
| 2,130,941,577
|
PR_kwDOJ0Z1Ps5mrdNr
| 2,468
|
Update llama.cpp submodule to `099afc6`
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-12T20:35:40
| 2024-02-12T22:01:17
| 2024-02-12T22:01:16
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2468",
"html_url": "https://github.com/ollama/ollama/pull/2468",
"diff_url": "https://github.com/ollama/ollama/pull/2468.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2468.patch",
"merged_at": "2024-02-12T22:01:16"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2468/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2468/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7450
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7450/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7450/comments
|
https://api.github.com/repos/ollama/ollama/issues/7450/events
|
https://github.com/ollama/ollama/issues/7450
| 2,627,248,628
|
I_kwDOJ0Z1Ps6cmKH0
| 7,450
|
Run LLM directly in Golang App without Ollama Server
|
{
"login": "faelp22",
"id": 6642575,
"node_id": "MDQ6VXNlcjY2NDI1NzU=",
"avatar_url": "https://avatars.githubusercontent.com/u/6642575?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/faelp22",
"html_url": "https://github.com/faelp22",
"followers_url": "https://api.github.com/users/faelp22/followers",
"following_url": "https://api.github.com/users/faelp22/following{/other_user}",
"gists_url": "https://api.github.com/users/faelp22/gists{/gist_id}",
"starred_url": "https://api.github.com/users/faelp22/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/faelp22/subscriptions",
"organizations_url": "https://api.github.com/users/faelp22/orgs",
"repos_url": "https://api.github.com/users/faelp22/repos",
"events_url": "https://api.github.com/users/faelp22/events{/privacy}",
"received_events_url": "https://api.github.com/users/faelp22/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 1
| 2024-10-31T16:04:31
| 2024-11-29T17:07:31
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello everyone, I would like to know if it is possible to run an all-minilm LLM model directly in my Golang App without having to make calls to the Ollama Server http://localhost:11434/api
I would like to take a small "all-minilm" model and use the //go:embed model/* to already have the tool embedded in the Golang binary.
The idea would be to simplify the use and infrastructure so as not to need to set up an extra Docker Container or VM, not to mention avoiding network latency.
I would like to use Ollama as a library and import the necessary parts in my Golang App.
I couldn't find any examples, if there is someone more experienced and can help me, I would appreciate it.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7450/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7450/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/7482
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7482/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7482/comments
|
https://api.github.com/repos/ollama/ollama/issues/7482/events
|
https://github.com/ollama/ollama/pull/7482
| 2,631,424,552
|
PR_kwDOJ0Z1Ps6Au4C6
| 7,482
|
Add action for publishing package to WinGet
|
{
"login": "mdanish-kh",
"id": 88161975,
"node_id": "MDQ6VXNlcjg4MTYxOTc1",
"avatar_url": "https://avatars.githubusercontent.com/u/88161975?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mdanish-kh",
"html_url": "https://github.com/mdanish-kh",
"followers_url": "https://api.github.com/users/mdanish-kh/followers",
"following_url": "https://api.github.com/users/mdanish-kh/following{/other_user}",
"gists_url": "https://api.github.com/users/mdanish-kh/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mdanish-kh/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mdanish-kh/subscriptions",
"organizations_url": "https://api.github.com/users/mdanish-kh/orgs",
"repos_url": "https://api.github.com/users/mdanish-kh/repos",
"events_url": "https://api.github.com/users/mdanish-kh/events{/privacy}",
"received_events_url": "https://api.github.com/users/mdanish-kh/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-11-03T19:38:14
| 2024-11-23T19:35:49
| 2024-11-23T19:35:48
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/7482",
"html_url": "https://github.com/ollama/ollama/pull/7482",
"diff_url": "https://github.com/ollama/ollama/pull/7482.diff",
"patch_url": "https://github.com/ollama/ollama/pull/7482.patch",
"merged_at": null
}
|
## Description
This PR proposes to add a GitHub action for submitting the latest stable release to WinGet as it gets published. [microsoft/winget-create](https://github.com/microsoft/winget-create) is used as the tool for submitting the latest package.
## Steps needed from maintainers
If the maintainers approve of these changes, they will need to do the following before merging this PR:
1. Fork [microsoft/winget-pkgs](https://github.com/microsoft/winget-pkgs) under a personal or bot account.
2. Create a [public access token (classic)](https://github.com/microsoft/winget-create?tab=readme-ov-file#github-personal-access-token-classic-permissions) with `public_repo` scope from the user account where the fork exists.
3. Create a repo secret in this repo with the name `WINGET_GITHUB_TOKEN`
For reference, maintainers may see similar implemented actions in the following repos:
[PowerToys](https://github.com/microsoft/PowerToys/blob/main/.github/workflows/package-submissions.yml), [Terminal](https://github.com/microsoft/terminal/blob/main/.github/workflows/winget.yml), [DevHome](https://github.com/microsoft/devhome/blob/main/.github/workflows/winget-submission.yml), [Oh-my-posh](https://github.com/JanDeDobbeleer/oh-my-posh/blob/main/.github/workflows/winget.yml)
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7482/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 1,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7482/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2520
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2520/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2520/comments
|
https://api.github.com/repos/ollama/ollama/issues/2520/events
|
https://github.com/ollama/ollama/issues/2520
| 2,137,351,508
|
I_kwDOJ0Z1Ps5_ZWVU
| 2,520
|
go-1.21 fails to build ollama: C source files not allowed when not using cgo or SWIG: gpu_info_cpu.c gpu_info_cuda.c gpu_info_rocm.c
|
{
"login": "yurivict",
"id": 271906,
"node_id": "MDQ6VXNlcjI3MTkwNg==",
"avatar_url": "https://avatars.githubusercontent.com/u/271906?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/yurivict",
"html_url": "https://github.com/yurivict",
"followers_url": "https://api.github.com/users/yurivict/followers",
"following_url": "https://api.github.com/users/yurivict/following{/other_user}",
"gists_url": "https://api.github.com/users/yurivict/gists{/gist_id}",
"starred_url": "https://api.github.com/users/yurivict/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/yurivict/subscriptions",
"organizations_url": "https://api.github.com/users/yurivict/orgs",
"repos_url": "https://api.github.com/users/yurivict/repos",
"events_url": "https://api.github.com/users/yurivict/events{/privacy}",
"received_events_url": "https://api.github.com/users/yurivict/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 11
| 2024-02-15T20:04:57
| 2024-05-02T22:00:23
| 2024-05-02T22:00:23
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
```
===> Building for ollama-0.1.25
(cd /usr/ports/misc/ollama/work/github.com/ollama/ollama@v0.1.25; for t in ./cmd; do out=$(/usr/bin/basename $(echo ${t} | /usr/bin/sed -Ee 's/^[^:]*:([^:]+).*$/\1/' -e 's/^\.$/ollama/')); pkg=$(echo ${t} | /usr/bin/sed -Ee 's/^([^:]*).*$/\1/' -e 's/^ollama$/./'); echo "===> Building ${out} from ${pkg}"; /usr/bin/env XDG_DATA_HOME=/usr/ports/misc/ollama/work XDG_CONFIG_HOME=/usr/ports/misc/ollama/work XDG_CACHE_HOME=/usr/ports/misc/ollama/work/.cache HOME=/usr/ports/misc/ollama/work PATH=/usr/local/libexec/ccache:/usr/ports/misc/ollama/work/.bin:/home/yuri/.cargo/bin:/home/yuri/bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin PKG_CONFIG_LIBDIR=/usr/ports/misc/ollama/work/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local CC="cc" CFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CPP="cpp" CPPFLAGS="" LDFLAGS=" -fstack-protector-strong " LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CCACHE_DIR="/tmp/.ccache" BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 0644" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" CGO_ENABLED=1 CGO_CFLAGS="-I/usr/local/include" CGO_LDFLAGS="-L/usr/local/lib" GOAMD64= GOARM= GOTMPDIR="/usr/ports/misc/ollama/work" GOPATH="/usr/ports/distfiles/go/misc_ollama" GOBIN="/usr/ports/misc/ollama/work/bin" GO111MODULE=on GOFLAGS=-modcacherw GOSUMDB=sum.golang.org GOMAXPROCS=7 GOPROXY=off /usr/local/bin/go121 build -buildmode=exe -v -trimpath -ldflags=-s -buildvcs=false -mod=vendor -o /usr/ports/misc/ollama/work/bin/${out} ${pkg}; done)
===> Building cmd from ./cmd
package github.com/jmorganca/ollama/cmd
imports github.com/jmorganca/ollama/server
imports github.com/jmorganca/ollama/gpu: C source files not allowed when not using cgo or SWIG: gpu_info_cpu.c gpu_info_cuda.c gpu_info_rocm.c
*** Error code 1
```
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2520/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2520/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1198
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1198/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1198/comments
|
https://api.github.com/repos/ollama/ollama/issues/1198/events
|
https://github.com/ollama/ollama/issues/1198
| 2,000,892,540
|
I_kwDOJ0Z1Ps53QzJ8
| 1,198
|
Support for hyenadna-large-1m-seqlen-hf
|
{
"login": "magedhelmy1",
"id": 63347261,
"node_id": "MDQ6VXNlcjYzMzQ3MjYx",
"avatar_url": "https://avatars.githubusercontent.com/u/63347261?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/magedhelmy1",
"html_url": "https://github.com/magedhelmy1",
"followers_url": "https://api.github.com/users/magedhelmy1/followers",
"following_url": "https://api.github.com/users/magedhelmy1/following{/other_user}",
"gists_url": "https://api.github.com/users/magedhelmy1/gists{/gist_id}",
"starred_url": "https://api.github.com/users/magedhelmy1/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/magedhelmy1/subscriptions",
"organizations_url": "https://api.github.com/users/magedhelmy1/orgs",
"repos_url": "https://api.github.com/users/magedhelmy1/repos",
"events_url": "https://api.github.com/users/magedhelmy1/events{/privacy}",
"received_events_url": "https://api.github.com/users/magedhelmy1/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
open
| false
| null |
[] | null | 1
| 2023-11-19T15:17:54
| 2024-03-11T17:46:16
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi, any plans to support hyenadna? it has 1 million tokens!
https://huggingface.co/LongSafari/hyenadna-large-1m-seqlen-hf
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1198/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1198/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/3234
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3234/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3234/comments
|
https://api.github.com/repos/ollama/ollama/issues/3234/events
|
https://github.com/ollama/ollama/issues/3234
| 2,193,959,556
|
I_kwDOJ0Z1Ps6CxSqE
| 3,234
|
is it possible to use ollama as a library , not through network
|
{
"login": "aizimuji",
"id": 129702132,
"node_id": "U_kgDOB7sY9A",
"avatar_url": "https://avatars.githubusercontent.com/u/129702132?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/aizimuji",
"html_url": "https://github.com/aizimuji",
"followers_url": "https://api.github.com/users/aizimuji/followers",
"following_url": "https://api.github.com/users/aizimuji/following{/other_user}",
"gists_url": "https://api.github.com/users/aizimuji/gists{/gist_id}",
"starred_url": "https://api.github.com/users/aizimuji/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/aizimuji/subscriptions",
"organizations_url": "https://api.github.com/users/aizimuji/orgs",
"repos_url": "https://api.github.com/users/aizimuji/repos",
"events_url": "https://api.github.com/users/aizimuji/events{/privacy}",
"received_events_url": "https://api.github.com/users/aizimuji/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] |
closed
| false
| null |
[] | null | 1
| 2024-03-19T03:58:11
| 2024-03-21T13:42:48
| 2024-03-21T13:42:47
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What are you trying to do?
i want to know if it's possible to call ollama function like some library, for example by dll in windows
so other developer can develop some gui or related software with this library
not to run ollama as a stand-alone server
### How should we solve this?
it's easier to build related gui or other third party software based on ollama
### What is the impact of not solving this?
_No response_
### Anything else?
_No response_
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3234/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3234/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2218
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2218/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2218/comments
|
https://api.github.com/repos/ollama/ollama/issues/2218/events
|
https://github.com/ollama/ollama/issues/2218
| 2,102,941,771
|
I_kwDOJ0Z1Ps59WFhL
| 2,218
|
:link: Please add HF (HuggingFace) model link to `duckdb-nsql` :duck:
|
{
"login": "adriens",
"id": 5235127,
"node_id": "MDQ6VXNlcjUyMzUxMjc=",
"avatar_url": "https://avatars.githubusercontent.com/u/5235127?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/adriens",
"html_url": "https://github.com/adriens",
"followers_url": "https://api.github.com/users/adriens/followers",
"following_url": "https://api.github.com/users/adriens/following{/other_user}",
"gists_url": "https://api.github.com/users/adriens/gists{/gist_id}",
"starred_url": "https://api.github.com/users/adriens/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/adriens/subscriptions",
"organizations_url": "https://api.github.com/users/adriens/orgs",
"repos_url": "https://api.github.com/users/adriens/repos",
"events_url": "https://api.github.com/users/adriens/events{/privacy}",
"received_events_url": "https://api.github.com/users/adriens/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 3
| 2024-01-26T21:40:29
| 2024-01-27T09:26:19
| 2024-01-27T06:25:50
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
# :grey_question: About
Recently, [`duckdb-nsql`](https://ollama.ai/library/duckdb-nsql) has been added to `ollama` library:
- https://github.com/ollama/ollama/issues/2193

**:point_right: ... but the page is lacking the HuggingFace model page.**
# :dart: Documentation request
Please add the following model [`motherduckdb/DuckDB-NSQL-7B-v0.1`](https://huggingface.co/motherduckdb/DuckDB-NSQL-7B-v0.1) url to [`duckdb-nsql` `ollama` page](https://ollama.ai/library/duckdb-nsql):
https://huggingface.co/motherduckdb/DuckDB-NSQL-7B-v0.1
# :moneybag: Benefits
- Better documentation
- Better indexation of `ollama` models
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2218/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2218/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6295
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6295/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6295/comments
|
https://api.github.com/repos/ollama/ollama/issues/6295/events
|
https://github.com/ollama/ollama/issues/6295
| 2,458,840,334
|
I_kwDOJ0Z1Ps6Sju0O
| 6,295
|
Ability to preload embedding model
|
{
"login": "comunidadio",
"id": 10286013,
"node_id": "MDQ6VXNlcjEwMjg2MDEz",
"avatar_url": "https://avatars.githubusercontent.com/u/10286013?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/comunidadio",
"html_url": "https://github.com/comunidadio",
"followers_url": "https://api.github.com/users/comunidadio/followers",
"following_url": "https://api.github.com/users/comunidadio/following{/other_user}",
"gists_url": "https://api.github.com/users/comunidadio/gists{/gist_id}",
"starred_url": "https://api.github.com/users/comunidadio/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/comunidadio/subscriptions",
"organizations_url": "https://api.github.com/users/comunidadio/orgs",
"repos_url": "https://api.github.com/users/comunidadio/repos",
"events_url": "https://api.github.com/users/comunidadio/events{/privacy}",
"received_events_url": "https://api.github.com/users/comunidadio/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 2
| 2024-08-10T01:30:27
| 2024-08-13T17:19:57
| 2024-08-13T17:19:57
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
The "empty request" trick to preload a model does not currently work for embedding models.
Source: https://github.com/ollama/ollama/blob/main/docs/faq.md#how-can-i-preload-a-model-into-ollama-to-get-faster-response-times and #2431
```
$ curl http://localhost:11434/api/embed -d '{"model": "all-minilm:latest"}'
{"error":"invalid input type"}%
```
```
ollama --version
ollama version is 0.3.4
```
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6295/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6295/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4748
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4748/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4748/comments
|
https://api.github.com/repos/ollama/ollama/issues/4748/events
|
https://github.com/ollama/ollama/issues/4748
| 2,327,501,621
|
I_kwDOJ0Z1Ps6Kuts1
| 4,748
|
Custom-llama issue
|
{
"login": "Ascariota",
"id": 25208125,
"node_id": "MDQ6VXNlcjI1MjA4MTI1",
"avatar_url": "https://avatars.githubusercontent.com/u/25208125?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Ascariota",
"html_url": "https://github.com/Ascariota",
"followers_url": "https://api.github.com/users/Ascariota/followers",
"following_url": "https://api.github.com/users/Ascariota/following{/other_user}",
"gists_url": "https://api.github.com/users/Ascariota/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Ascariota/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Ascariota/subscriptions",
"organizations_url": "https://api.github.com/users/Ascariota/orgs",
"repos_url": "https://api.github.com/users/Ascariota/repos",
"events_url": "https://api.github.com/users/Ascariota/events{/privacy}",
"received_events_url": "https://api.github.com/users/Ascariota/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 0
| 2024-05-31T10:29:12
| 2024-05-31T10:29:12
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hello,
I have a problem, or I misunderstood something. If I put several SYSTEM tags in my custom-llama3 file, only the last one is taken.
How can I give him more information?
Example I would like
SYSTEM You are a helpful AI assistant named Droid
but also that they can know the location of the smoking area in my building, or that to contact the Sales department they must contact 00*******?
thank you in advance for your help and sorry if my question is stupid.
### OS
Windows
### GPU
Other
### CPU
Intel
### Ollama version
0.1.39
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4748/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4748/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/6977
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6977/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6977/comments
|
https://api.github.com/repos/ollama/ollama/issues/6977/events
|
https://github.com/ollama/ollama/issues/6977
| 2,549,983,967
|
I_kwDOJ0Z1Ps6X_arf
| 6,977
|
To configure Ollama to run multiple models simultaneously
|
{
"login": "DavidAlpha007",
"id": 143383189,
"node_id": "U_kgDOCIvalQ",
"avatar_url": "https://avatars.githubusercontent.com/u/143383189?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/DavidAlpha007",
"html_url": "https://github.com/DavidAlpha007",
"followers_url": "https://api.github.com/users/DavidAlpha007/followers",
"following_url": "https://api.github.com/users/DavidAlpha007/following{/other_user}",
"gists_url": "https://api.github.com/users/DavidAlpha007/gists{/gist_id}",
"starred_url": "https://api.github.com/users/DavidAlpha007/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/DavidAlpha007/subscriptions",
"organizations_url": "https://api.github.com/users/DavidAlpha007/orgs",
"repos_url": "https://api.github.com/users/DavidAlpha007/repos",
"events_url": "https://api.github.com/users/DavidAlpha007/events{/privacy}",
"received_events_url": "https://api.github.com/users/DavidAlpha007/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396220,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2afA",
"url": "https://api.github.com/repos/ollama/ollama/labels/question",
"name": "question",
"color": "d876e3",
"default": true,
"description": "General questions"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 2
| 2024-09-26T09:08:05
| 2024-09-26T15:46:40
| 2024-09-26T15:46:36
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
if the design of Ollama can support calling multiple models simultaneously. For example, can it be used in evaluation scenarios? Thanks for your support.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6977/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6977/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1102
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1102/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1102/comments
|
https://api.github.com/repos/ollama/ollama/issues/1102/events
|
https://github.com/ollama/ollama/issues/1102
| 1,989,549,140
|
I_kwDOJ0Z1Ps52lhxU
| 1,102
|
Ollama on FreeBSD
|
{
"login": "eng-alameedi",
"id": 73557986,
"node_id": "MDQ6VXNlcjczNTU3OTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/73557986?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/eng-alameedi",
"html_url": "https://github.com/eng-alameedi",
"followers_url": "https://api.github.com/users/eng-alameedi/followers",
"following_url": "https://api.github.com/users/eng-alameedi/following{/other_user}",
"gists_url": "https://api.github.com/users/eng-alameedi/gists{/gist_id}",
"starred_url": "https://api.github.com/users/eng-alameedi/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/eng-alameedi/subscriptions",
"organizations_url": "https://api.github.com/users/eng-alameedi/orgs",
"repos_url": "https://api.github.com/users/eng-alameedi/repos",
"events_url": "https://api.github.com/users/eng-alameedi/events{/privacy}",
"received_events_url": "https://api.github.com/users/eng-alameedi/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 46
| 2023-11-12T20:07:58
| 2024-11-08T22:12:37
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello there:
is there any chance to get ollama working on freebsd please??
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1102/reactions",
"total_count": 4,
"+1": 4,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1102/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/2460
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2460/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2460/comments
|
https://api.github.com/repos/ollama/ollama/issues/2460/events
|
https://github.com/ollama/ollama/pull/2460
| 2,129,533,959
|
PR_kwDOJ0Z1Ps5mmnoO
| 2,460
|
Refactor chat prompt templating
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-12T07:03:05
| 2024-02-12T23:06:58
| 2024-02-12T23:06:57
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2460",
"html_url": "https://github.com/ollama/ollama/pull/2460",
"diff_url": "https://github.com/ollama/ollama/pull/2460.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2460.patch",
"merged_at": "2024-02-12T23:06:57"
}
|
This refactors the chat prompt processing to be a little easier to follow. It also fully deprecates `.First` in favor of the chat endpoint
Fixes https://github.com/ollama/ollama/issues/2443
Fixes https://github.com/ollama/ollama/issues/2438
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2460/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2460/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/7391
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7391/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7391/comments
|
https://api.github.com/repos/ollama/ollama/issues/7391/events
|
https://github.com/ollama/ollama/issues/7391
| 2,617,261,796
|
I_kwDOJ0Z1Ps6cAD7k
| 7,391
|
ollama -v return 2version one is 0.0.0 the other is client version 0.3.14
|
{
"login": "FanGShiYuu",
"id": 88468647,
"node_id": "MDQ6VXNlcjg4NDY4NjQ3",
"avatar_url": "https://avatars.githubusercontent.com/u/88468647?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/FanGShiYuu",
"html_url": "https://github.com/FanGShiYuu",
"followers_url": "https://api.github.com/users/FanGShiYuu/followers",
"following_url": "https://api.github.com/users/FanGShiYuu/following{/other_user}",
"gists_url": "https://api.github.com/users/FanGShiYuu/gists{/gist_id}",
"starred_url": "https://api.github.com/users/FanGShiYuu/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/FanGShiYuu/subscriptions",
"organizations_url": "https://api.github.com/users/FanGShiYuu/orgs",
"repos_url": "https://api.github.com/users/FanGShiYuu/repos",
"events_url": "https://api.github.com/users/FanGShiYuu/events{/privacy}",
"received_events_url": "https://api.github.com/users/FanGShiYuu/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-10-28T04:25:23
| 2024-11-04T17:59:44
| 2024-11-04T17:59:44
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
i am using ubuntu20.04;install ollama through curl -fsSL https://ollama.com/install.sh | sh; when input ollama -v return ollama version is 0.0.0
Warning: client version is 0.3.14
btw, when using ollama, i notice my gpu is not used and the response is so slow
### OS
Linux
### GPU
Nvidia
### CPU
_No response_
### Ollama version
0.3.14
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7391/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7391/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2959
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2959/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2959/comments
|
https://api.github.com/repos/ollama/ollama/issues/2959/events
|
https://github.com/ollama/ollama/pull/2959
| 2,172,301,959
|
PR_kwDOJ0Z1Ps5o4eiY
| 2,959
|
fix json encoder
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-03-06T19:48:35
| 2024-05-09T22:18:42
| 2024-03-06T21:04:14
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2959",
"html_url": "https://github.com/ollama/ollama/pull/2959",
"diff_url": "https://github.com/ollama/ollama/pull/2959.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2959.patch",
"merged_at": null
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2959/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2959/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8201
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8201/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8201/comments
|
https://api.github.com/repos/ollama/ollama/issues/8201/events
|
https://github.com/ollama/ollama/issues/8201
| 2,754,242,988
|
I_kwDOJ0Z1Ps6kKmms
| 8,201
|
Ollama
|
{
"login": "Sandro127",
"id": 149949677,
"node_id": "U_kgDOCPAM7Q",
"avatar_url": "https://avatars.githubusercontent.com/u/149949677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Sandro127",
"html_url": "https://github.com/Sandro127",
"followers_url": "https://api.github.com/users/Sandro127/followers",
"following_url": "https://api.github.com/users/Sandro127/following{/other_user}",
"gists_url": "https://api.github.com/users/Sandro127/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Sandro127/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Sandro127/subscriptions",
"organizations_url": "https://api.github.com/users/Sandro127/orgs",
"repos_url": "https://api.github.com/users/Sandro127/repos",
"events_url": "https://api.github.com/users/Sandro127/events{/privacy}",
"received_events_url": "https://api.github.com/users/Sandro127/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-12-21T16:57:55
| 2024-12-21T16:58:10
| 2024-12-21T16:58:10
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null | null |
{
"login": "Sandro127",
"id": 149949677,
"node_id": "U_kgDOCPAM7Q",
"avatar_url": "https://avatars.githubusercontent.com/u/149949677?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Sandro127",
"html_url": "https://github.com/Sandro127",
"followers_url": "https://api.github.com/users/Sandro127/followers",
"following_url": "https://api.github.com/users/Sandro127/following{/other_user}",
"gists_url": "https://api.github.com/users/Sandro127/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Sandro127/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Sandro127/subscriptions",
"organizations_url": "https://api.github.com/users/Sandro127/orgs",
"repos_url": "https://api.github.com/users/Sandro127/repos",
"events_url": "https://api.github.com/users/Sandro127/events{/privacy}",
"received_events_url": "https://api.github.com/users/Sandro127/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8201/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8201/timeline
| null |
not_planned
| false
|
https://api.github.com/repos/ollama/ollama/issues/166
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/166/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/166/comments
|
https://api.github.com/repos/ollama/ollama/issues/166/events
|
https://github.com/ollama/ollama/pull/166
| 1,816,357,300
|
PR_kwDOJ0Z1Ps5WIKOs
| 166
|
Note that CGO must be enabled in dev docs
|
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-21T20:36:39
| 2023-07-21T20:48:17
| 2023-07-21T20:48:10
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/166",
"html_url": "https://github.com/ollama/ollama/pull/166",
"diff_url": "https://github.com/ollama/ollama/pull/166.diff",
"patch_url": "https://github.com/ollama/ollama/pull/166.patch",
"merged_at": "2023-07-21T20:48:10"
}
| null |
{
"login": "BruceMacD",
"id": 5853428,
"node_id": "MDQ6VXNlcjU4NTM0Mjg=",
"avatar_url": "https://avatars.githubusercontent.com/u/5853428?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/BruceMacD",
"html_url": "https://github.com/BruceMacD",
"followers_url": "https://api.github.com/users/BruceMacD/followers",
"following_url": "https://api.github.com/users/BruceMacD/following{/other_user}",
"gists_url": "https://api.github.com/users/BruceMacD/gists{/gist_id}",
"starred_url": "https://api.github.com/users/BruceMacD/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/BruceMacD/subscriptions",
"organizations_url": "https://api.github.com/users/BruceMacD/orgs",
"repos_url": "https://api.github.com/users/BruceMacD/repos",
"events_url": "https://api.github.com/users/BruceMacD/events{/privacy}",
"received_events_url": "https://api.github.com/users/BruceMacD/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/166/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/166/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8310
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8310/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8310/comments
|
https://api.github.com/repos/ollama/ollama/issues/8310/events
|
https://github.com/ollama/ollama/issues/8310
| 2,769,334,462
|
I_kwDOJ0Z1Ps6lELC-
| 8,310
|
llama3.2-vision doesn't utilize my GPU.
|
{
"login": "blueApple12",
"id": 89522107,
"node_id": "MDQ6VXNlcjg5NTIyMTA3",
"avatar_url": "https://avatars.githubusercontent.com/u/89522107?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/blueApple12",
"html_url": "https://github.com/blueApple12",
"followers_url": "https://api.github.com/users/blueApple12/followers",
"following_url": "https://api.github.com/users/blueApple12/following{/other_user}",
"gists_url": "https://api.github.com/users/blueApple12/gists{/gist_id}",
"starred_url": "https://api.github.com/users/blueApple12/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/blueApple12/subscriptions",
"organizations_url": "https://api.github.com/users/blueApple12/orgs",
"repos_url": "https://api.github.com/users/blueApple12/repos",
"events_url": "https://api.github.com/users/blueApple12/events{/privacy}",
"received_events_url": "https://api.github.com/users/blueApple12/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 32
| 2025-01-05T15:51:49
| 2025-01-17T18:30:18
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I bought a new pc with 4070 Super to do some AI tasks using Ollama, but when I tried to run llama3.2-vision it just didn't utilize my GPU and only utilize my CPU, llama3.2 does utilize my GPU, so why is that? thank you.
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.5.4
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8310/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8310/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/175
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/175/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/175/comments
|
https://api.github.com/repos/ollama/ollama/issues/175/events
|
https://github.com/ollama/ollama/pull/175
| 1,816,776,380
|
PR_kwDOJ0Z1Ps5WJfEd
| 175
|
Update .gitignore
|
{
"login": "jk1jk",
"id": 140257749,
"node_id": "U_kgDOCFwp1Q",
"avatar_url": "https://avatars.githubusercontent.com/u/140257749?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jk1jk",
"html_url": "https://github.com/jk1jk",
"followers_url": "https://api.github.com/users/jk1jk/followers",
"following_url": "https://api.github.com/users/jk1jk/following{/other_user}",
"gists_url": "https://api.github.com/users/jk1jk/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jk1jk/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jk1jk/subscriptions",
"organizations_url": "https://api.github.com/users/jk1jk/orgs",
"repos_url": "https://api.github.com/users/jk1jk/repos",
"events_url": "https://api.github.com/users/jk1jk/events{/privacy}",
"received_events_url": "https://api.github.com/users/jk1jk/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-07-22T14:03:26
| 2023-07-22T16:40:38
| 2023-07-22T16:40:38
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/175",
"html_url": "https://github.com/ollama/ollama/pull/175",
"diff_url": "https://github.com/ollama/ollama/pull/175.diff",
"patch_url": "https://github.com/ollama/ollama/pull/175.patch",
"merged_at": "2023-07-22T16:40:38"
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/175/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/175/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2403
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2403/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2403/comments
|
https://api.github.com/repos/ollama/ollama/issues/2403/events
|
https://github.com/ollama/ollama/pull/2403
| 2,124,200,380
|
PR_kwDOJ0Z1Ps5mUxPw
| 2,403
|
Ensure the libraries are present
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2024-02-08T01:28:22
| 2024-02-08T01:55:33
| 2024-02-08T01:55:31
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2403",
"html_url": "https://github.com/ollama/ollama/pull/2403",
"diff_url": "https://github.com/ollama/ollama/pull/2403.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2403.patch",
"merged_at": "2024-02-08T01:55:31"
}
|
When we store our libraries in a temp dir, a reaper might clean them when we are idle, so make sure to check for them before we reload.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2403/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2403/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/3579
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3579/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3579/comments
|
https://api.github.com/repos/ollama/ollama/issues/3579/events
|
https://github.com/ollama/ollama/pull/3579
| 2,236,185,189
|
PR_kwDOJ0Z1Ps5sRyQc
| 3,579
|
fix ci
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-04-10T18:27:10
| 2024-04-10T18:37:02
| 2024-04-10T18:37:01
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3579",
"html_url": "https://github.com/ollama/ollama/pull/3579",
"diff_url": "https://github.com/ollama/ollama/pull/3579.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3579.patch",
"merged_at": "2024-04-10T18:37:01"
}
| null |
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3579/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3579/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/675
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/675/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/675/comments
|
https://api.github.com/repos/ollama/ollama/issues/675/events
|
https://github.com/ollama/ollama/issues/675
| 1,922,472,936
|
I_kwDOJ0Z1Ps5ylpvo
| 675
|
api improvements
|
{
"login": "jtoy",
"id": 14783,
"node_id": "MDQ6VXNlcjE0Nzgz",
"avatar_url": "https://avatars.githubusercontent.com/u/14783?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jtoy",
"html_url": "https://github.com/jtoy",
"followers_url": "https://api.github.com/users/jtoy/followers",
"following_url": "https://api.github.com/users/jtoy/following{/other_user}",
"gists_url": "https://api.github.com/users/jtoy/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jtoy/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jtoy/subscriptions",
"organizations_url": "https://api.github.com/users/jtoy/orgs",
"repos_url": "https://api.github.com/users/jtoy/repos",
"events_url": "https://api.github.com/users/jtoy/events{/privacy}",
"received_events_url": "https://api.github.com/users/jtoy/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 7
| 2023-10-02T18:59:10
| 2024-01-10T13:14:47
| 2023-10-05T16:38:04
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
its a stream of objects that are separated with a newline. often times new lines are returned in the response, so that breaks just splitting on new lines.
I think the split should be on something else.
Also it seems like there should be an api endpoint that just returns the whole response in a string.
thoughts?
|
{
"login": "technovangelist",
"id": 633681,
"node_id": "MDQ6VXNlcjYzMzY4MQ==",
"avatar_url": "https://avatars.githubusercontent.com/u/633681?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/technovangelist",
"html_url": "https://github.com/technovangelist",
"followers_url": "https://api.github.com/users/technovangelist/followers",
"following_url": "https://api.github.com/users/technovangelist/following{/other_user}",
"gists_url": "https://api.github.com/users/technovangelist/gists{/gist_id}",
"starred_url": "https://api.github.com/users/technovangelist/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/technovangelist/subscriptions",
"organizations_url": "https://api.github.com/users/technovangelist/orgs",
"repos_url": "https://api.github.com/users/technovangelist/repos",
"events_url": "https://api.github.com/users/technovangelist/events{/privacy}",
"received_events_url": "https://api.github.com/users/technovangelist/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/675/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/675/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/3149
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3149/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3149/comments
|
https://api.github.com/repos/ollama/ollama/issues/3149/events
|
https://github.com/ollama/ollama/pull/3149
| 2,187,149,183
|
PR_kwDOJ0Z1Ps5prGX9
| 3,149
|
fix: clip memory leak
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-03-14T19:47:41
| 2024-03-14T20:34:16
| 2024-03-14T20:34:15
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/3149",
"html_url": "https://github.com/ollama/ollama/pull/3149",
"diff_url": "https://github.com/ollama/ollama/pull/3149.diff",
"patch_url": "https://github.com/ollama/ollama/pull/3149.patch",
"merged_at": "2024-03-14T20:34:15"
}
|
this change patches llama.cpp and fixes two bugs
1. llama_server_context never calls clip_free
2. clip_free does not fully free its context
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3149/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3149/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1212
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1212/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1212/comments
|
https://api.github.com/repos/ollama/ollama/issues/1212/events
|
https://github.com/ollama/ollama/pull/1212
| 2,003,065,238
|
PR_kwDOJ0Z1Ps5f9NN9
| 1,212
|
enable metal for fp32, q5_0, q5_1
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-11-20T21:48:29
| 2023-11-20T21:56:41
| 2023-11-20T21:56:40
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1212",
"html_url": "https://github.com/ollama/ollama/pull/1212",
"diff_url": "https://github.com/ollama/ollama/pull/1212.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1212.patch",
"merged_at": "2023-11-20T21:56:40"
}
|
recent llama.cpp update added kernels for fp32, q5_0, and q5_1
resolves #1200
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1212/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 1,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1212/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8285
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8285/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8285/comments
|
https://api.github.com/repos/ollama/ollama/issues/8285/events
|
https://github.com/ollama/ollama/issues/8285
| 2,765,781,740
|
I_kwDOJ0Z1Ps6k2nrs
| 8,285
|
GPU runs at maximum load with 2 models
|
{
"login": "RomanDrechsel",
"id": 6135586,
"node_id": "MDQ6VXNlcjYxMzU1ODY=",
"avatar_url": "https://avatars.githubusercontent.com/u/6135586?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/RomanDrechsel",
"html_url": "https://github.com/RomanDrechsel",
"followers_url": "https://api.github.com/users/RomanDrechsel/followers",
"following_url": "https://api.github.com/users/RomanDrechsel/following{/other_user}",
"gists_url": "https://api.github.com/users/RomanDrechsel/gists{/gist_id}",
"starred_url": "https://api.github.com/users/RomanDrechsel/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/RomanDrechsel/subscriptions",
"organizations_url": "https://api.github.com/users/RomanDrechsel/orgs",
"repos_url": "https://api.github.com/users/RomanDrechsel/repos",
"events_url": "https://api.github.com/users/RomanDrechsel/events{/privacy}",
"received_events_url": "https://api.github.com/users/RomanDrechsel/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 21
| 2025-01-02T10:27:48
| 2025-01-24T21:56:15
| 2025-01-11T06:51:46
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
Hi,
i use ollama as the provider for the Continue extension for VSCode for tab autocompletion.
Since the last update I have the problem that my GPU runs at maximum load as soon as 2 models are running at the same time.
Even if they are only very small models (e.g. nomic-embed-text for embeddings and qwen2.5-coder:0.5b as tab autocomplete).
The load remains at 100% until I stop one of the two models.
Before the last update, I had no problems using larger models (e.g. qwen2.5-coder:3b).
My OS is Manjaro Linux with kernel 6.12.4,
my hardware is an AMD Ryzen 9 9950X processor and an AMD Radeon RX 7900 XTX.
### OS
Linux
### GPU
AMD
### CPU
AMD
### Ollama version
0.5.4
|
{
"login": "rick-github",
"id": 14946854,
"node_id": "MDQ6VXNlcjE0OTQ2ODU0",
"avatar_url": "https://avatars.githubusercontent.com/u/14946854?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/rick-github",
"html_url": "https://github.com/rick-github",
"followers_url": "https://api.github.com/users/rick-github/followers",
"following_url": "https://api.github.com/users/rick-github/following{/other_user}",
"gists_url": "https://api.github.com/users/rick-github/gists{/gist_id}",
"starred_url": "https://api.github.com/users/rick-github/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/rick-github/subscriptions",
"organizations_url": "https://api.github.com/users/rick-github/orgs",
"repos_url": "https://api.github.com/users/rick-github/repos",
"events_url": "https://api.github.com/users/rick-github/events{/privacy}",
"received_events_url": "https://api.github.com/users/rick-github/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8285/reactions",
"total_count": 1,
"+1": 1,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8285/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/5397
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5397/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5397/comments
|
https://api.github.com/repos/ollama/ollama/issues/5397/events
|
https://github.com/ollama/ollama/issues/5397
| 2,382,728,555
|
I_kwDOJ0Z1Ps6OBY1r
| 5,397
|
V0.1.48 The model is loaded into the GPU Memory but runs on the CPU
|
{
"login": "wxtt522",
"id": 28422636,
"node_id": "MDQ6VXNlcjI4NDIyNjM2",
"avatar_url": "https://avatars.githubusercontent.com/u/28422636?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wxtt522",
"html_url": "https://github.com/wxtt522",
"followers_url": "https://api.github.com/users/wxtt522/followers",
"following_url": "https://api.github.com/users/wxtt522/following{/other_user}",
"gists_url": "https://api.github.com/users/wxtt522/gists{/gist_id}",
"starred_url": "https://api.github.com/users/wxtt522/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/wxtt522/subscriptions",
"organizations_url": "https://api.github.com/users/wxtt522/orgs",
"repos_url": "https://api.github.com/users/wxtt522/repos",
"events_url": "https://api.github.com/users/wxtt522/events{/privacy}",
"received_events_url": "https://api.github.com/users/wxtt522/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 5860134234,
"node_id": "LA_kwDOJ0Z1Ps8AAAABXUqNWg",
"url": "https://api.github.com/repos/ollama/ollama/labels/windows",
"name": "windows",
"color": "0052CC",
"default": false,
"description": ""
},
{
"id": 6677367769,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgCL2Q",
"url": "https://api.github.com/repos/ollama/ollama/labels/needs%20more%20info",
"name": "needs more info",
"color": "BA8041",
"default": false,
"description": "More information is needed to assist"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 5
| 2024-07-01T03:38:43
| 2024-07-03T07:26:48
| 2024-07-03T07:26:48
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
ollama run gemma2:27b

The same goes for loading other models. It was normal in the previous version. I did not change any environment variables.
### OS
Windows
### GPU
Nvidia
### CPU
Intel
### Ollama version
0.1.48
|
{
"login": "wxtt522",
"id": 28422636,
"node_id": "MDQ6VXNlcjI4NDIyNjM2",
"avatar_url": "https://avatars.githubusercontent.com/u/28422636?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/wxtt522",
"html_url": "https://github.com/wxtt522",
"followers_url": "https://api.github.com/users/wxtt522/followers",
"following_url": "https://api.github.com/users/wxtt522/following{/other_user}",
"gists_url": "https://api.github.com/users/wxtt522/gists{/gist_id}",
"starred_url": "https://api.github.com/users/wxtt522/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/wxtt522/subscriptions",
"organizations_url": "https://api.github.com/users/wxtt522/orgs",
"repos_url": "https://api.github.com/users/wxtt522/repos",
"events_url": "https://api.github.com/users/wxtt522/events{/privacy}",
"received_events_url": "https://api.github.com/users/wxtt522/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5397/reactions",
"total_count": 3,
"+1": 3,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5397/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/656
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/656/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/656/comments
|
https://api.github.com/repos/ollama/ollama/issues/656/events
|
https://github.com/ollama/ollama/issues/656
| 1,920,166,650
|
I_kwDOJ0Z1Ps5yc2r6
| 656
|
CLI run output not standard output
|
{
"login": "reustle",
"id": 304560,
"node_id": "MDQ6VXNlcjMwNDU2MA==",
"avatar_url": "https://avatars.githubusercontent.com/u/304560?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/reustle",
"html_url": "https://github.com/reustle",
"followers_url": "https://api.github.com/users/reustle/followers",
"following_url": "https://api.github.com/users/reustle/following{/other_user}",
"gists_url": "https://api.github.com/users/reustle/gists{/gist_id}",
"starred_url": "https://api.github.com/users/reustle/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/reustle/subscriptions",
"organizations_url": "https://api.github.com/users/reustle/orgs",
"repos_url": "https://api.github.com/users/reustle/repos",
"events_url": "https://api.github.com/users/reustle/events{/privacy}",
"received_events_url": "https://api.github.com/users/reustle/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 4
| 2023-09-30T08:04:16
| 2023-10-02T18:52:16
| 2023-10-02T18:52:16
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hello, I've been on this for quite some time now, and I'm sorry if I'm misinformed.
To me, it seems like even when I use the command line argument style input such as `ollama run mistral "Here is my prompt"` (as mentioned here https://github.com/jmorganca/ollama#pass-in-prompt-as-arguments ), the output isn't clean text.
When I run that command manually, while it should be just straight text with newline characters, instead it is doing some other characters to always fit the width of the terminal that called the command. Here's an example of `ollama run mistral "Here is my prompt" > out.txt`. It is adding some strange characters in the output.
<img width="256" alt="image" src="https://github.com/jmorganca/ollama/assets/304560/6e8152e9-b5c8-448c-9d7c-9d2fc8661924">
I think this has to do with how Ollama handles terminal commands, similar to when you use interactive chat mode. I would expect that the little loading ascii icon should not show when I'm using it as a standard command line tool (passing in the prompt directly vs chat mode).
If my understanding is correct here, and you're in agreement that we shouldn't be using the fancy terminal features and instead just outputing as STDOUT when it is finished processing, I'm happy to take a swing at creating a PR to fix this case.
Thank you!
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/656/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/656/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/1986
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1986/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1986/comments
|
https://api.github.com/repos/ollama/ollama/issues/1986/events
|
https://github.com/ollama/ollama/issues/1986
| 2,080,614,336
|
I_kwDOJ0Z1Ps58A6fA
| 1,986
|
Ollama Utilizing Only CPU Instead of GPU on MacBook Pro M1 Pro
|
{
"login": "vidvudsc",
"id": 77242455,
"node_id": "MDQ6VXNlcjc3MjQyNDU1",
"avatar_url": "https://avatars.githubusercontent.com/u/77242455?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/vidvudsc",
"html_url": "https://github.com/vidvudsc",
"followers_url": "https://api.github.com/users/vidvudsc/followers",
"following_url": "https://api.github.com/users/vidvudsc/following{/other_user}",
"gists_url": "https://api.github.com/users/vidvudsc/gists{/gist_id}",
"starred_url": "https://api.github.com/users/vidvudsc/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/vidvudsc/subscriptions",
"organizations_url": "https://api.github.com/users/vidvudsc/orgs",
"repos_url": "https://api.github.com/users/vidvudsc/repos",
"events_url": "https://api.github.com/users/vidvudsc/events{/privacy}",
"received_events_url": "https://api.github.com/users/vidvudsc/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 9
| 2024-01-14T07:18:33
| 2024-06-29T17:51:50
| 2024-01-14T19:14:14
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Description
I've encountered an issue where Ollama, when running any llm is utilizing only the CPU instead of the GPU on my MacBook Pro with an M1 Pro chip. This results in less efficient model performance than expected.
Environment
MacBook Pro with M1 Pro chip
MacOS version: Sonoma 14.2.1
Ollama version: 1.20
No specific error messages are observed.
All dependencies and drivers are up to date.
I would appreciate any guidance or updates regarding this issue. If there are any configurations or settings I might be missing, please let me know.
<img width="852" alt="Screenshot 2024-01-14 at 08 00 10" src="https://github.com/jmorganca/ollama/assets/77242455/ee3c0398-37e9-4473-af5b-a3b3253d1662">
PS. the image was taken when running dolphin-mixtral
Thanks!
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1986/reactions",
"total_count": 2,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 2
}
|
https://api.github.com/repos/ollama/ollama/issues/1986/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/36
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/36/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/36/comments
|
https://api.github.com/repos/ollama/ollama/issues/36/events
|
https://github.com/ollama/ollama/issues/36
| 1,786,490,994
|
I_kwDOJ0Z1Ps5qe7By
| 36
|
Fetch `q4_k` models from hugging face
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 0
| 2023-07-03T16:25:37
| 2023-07-08T03:26:50
| 2023-07-08T03:26:50
|
MEMBER
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
On macOS, metal only supports 4-bit and 16-bit quantization
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/36/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/36/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6173
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6173/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6173/comments
|
https://api.github.com/repos/ollama/ollama/issues/6173/events
|
https://github.com/ollama/ollama/issues/6173
| 2,447,913,094
|
I_kwDOJ0Z1Ps6R6DCG
| 6,173
|
Using ollama version 0.3.3, downloading all models will result in errors.
|
{
"login": "ucjmhfeng",
"id": 65010234,
"node_id": "MDQ6VXNlcjY1MDEwMjM0",
"avatar_url": "https://avatars.githubusercontent.com/u/65010234?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ucjmhfeng",
"html_url": "https://github.com/ucjmhfeng",
"followers_url": "https://api.github.com/users/ucjmhfeng/followers",
"following_url": "https://api.github.com/users/ucjmhfeng/following{/other_user}",
"gists_url": "https://api.github.com/users/ucjmhfeng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ucjmhfeng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ucjmhfeng/subscriptions",
"organizations_url": "https://api.github.com/users/ucjmhfeng/orgs",
"repos_url": "https://api.github.com/users/ucjmhfeng/repos",
"events_url": "https://api.github.com/users/ucjmhfeng/events{/privacy}",
"received_events_url": "https://api.github.com/users/ucjmhfeng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 7
| 2024-08-05T08:17:10
| 2024-08-30T12:32:54
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
pulling manifest
Error: pull model manifest: Get "https://registry.ollama.ai/v2/library/gemma2/manifests/2b": net/http: TLS handshake timeout.
Before version 0.3.0, there were no similar issues. Starting from the update to 0.3.1, I tried many methods, but none of them worked, including using VPN and reinstalling Olama.
### OS
Linux
### GPU
Other
### CPU
Intel
### Ollama version
0.3.3
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6173/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6173/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/5706
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5706/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5706/comments
|
https://api.github.com/repos/ollama/ollama/issues/5706/events
|
https://github.com/ollama/ollama/issues/5706
| 2,409,283,935
|
I_kwDOJ0Z1Ps6PmsFf
| 5,706
|
Multiple windows instances with different ports
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1
| 2024-07-15T17:44:03
| 2024-07-16T02:51:39
| null |
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
When you set an alternate port for OLLAMA_HOST, the CLI will spawn a new app, and create multiple tray instances that have no way to tell which one represents which port.
### OS
Windows
### GPU
_No response_
### CPU
_No response_
### Ollama version
_No response_
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5706/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5706/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/4965
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4965/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4965/comments
|
https://api.github.com/repos/ollama/ollama/issues/4965/events
|
https://github.com/ollama/ollama/pull/4965
| 2,344,553,985
|
PR_kwDOJ0Z1Ps5yAx3F
| 4,965
|
fix: skip removing layers that no longer exist
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-06-10T18:18:33
| 2024-06-10T18:40:04
| 2024-06-10T18:40:03
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/4965",
"html_url": "https://github.com/ollama/ollama/pull/4965",
"diff_url": "https://github.com/ollama/ollama/pull/4965.diff",
"patch_url": "https://github.com/ollama/ollama/pull/4965.patch",
"merged_at": "2024-06-10T18:40:03"
}
|
some models, such as `wizardcoder:34b-python`, incorrectly includes the config layer as an item in layers. this causes `RemoveLayers` to try to remove the same layer more than once, failing the second time since it's already removed
```json
{
"schemaVersion": 2,
"mediaType": "application/vnd.docker.distribution.manifest.v2+json",
"config": {
"mediaType": "application/vnd.docker.container.image.v1+json",
"digest": "sha256:a168bedb9a09640289c5174690a6221adae48b75dc431a219923f052ef20d0af",
"size": 456
},
"layers": [
{
"mediaType": "application/vnd.ollama.image.model",
"digest": "sha256:c8ad30822293b3c24f02265a42e2879b9725bd19a54d048fbe2c38487ee0ea84",
"size": 19052059872
},
{
"mediaType": "application/vnd.ollama.image.template",
"digest": "sha256:2d836d77287d85ac3d2ea87f4d765db6aaabc98543442072111b3d9831cdf9f1",
"size": 61
},
{
"mediaType": "application/vnd.ollama.image.system",
"digest": "sha256:8fadb9ad1206f2f130b54004893a2a7f76b1ff41a78049d69d797df2ee67fe94",
"size": 106
},
{
"mediaType": "application/vnd.ollama.image.params",
"digest": "sha256:bf6237d30a42b25b196a7a178dc566e113cf2f193aa11e7302c6d61880be6028",
"size": 30
},
{
"mediaType": "application/vnd.docker.container.image.v1+json",
"digest": "sha256:a168bedb9a09640289c5174690a6221adae48b75dc431a219923f052ef20d0af",
"size": 456
}
]
}
```
resolves #4898
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4965/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4965/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/8491
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/8491/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/8491/comments
|
https://api.github.com/repos/ollama/ollama/issues/8491/events
|
https://github.com/ollama/ollama/issues/8491
| 2,797,924,474
|
I_kwDOJ0Z1Ps6mxPB6
| 8,491
|
MacApp fails to build when building from source
|
{
"login": "devlux76",
"id": 86517969,
"node_id": "MDQ6VXNlcjg2NTE3OTY5",
"avatar_url": "https://avatars.githubusercontent.com/u/86517969?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/devlux76",
"html_url": "https://github.com/devlux76",
"followers_url": "https://api.github.com/users/devlux76/followers",
"following_url": "https://api.github.com/users/devlux76/following{/other_user}",
"gists_url": "https://api.github.com/users/devlux76/gists{/gist_id}",
"starred_url": "https://api.github.com/users/devlux76/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/devlux76/subscriptions",
"organizations_url": "https://api.github.com/users/devlux76/orgs",
"repos_url": "https://api.github.com/users/devlux76/repos",
"events_url": "https://api.github.com/users/devlux76/events{/privacy}",
"received_events_url": "https://api.github.com/users/devlux76/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 0
| 2025-01-20T00:32:59
| 2025-01-20T00:33:36
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I cloned the repo and was building the macapp and it fails to build. Can't find webpack.main.config
There's a webpack.main.config.ts file but that's not the file referenced. I tried to fix it myself and fell down a rabbit hole.
I'm just bringing this to the attention of whomever is maintaining it.
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
current head
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/8491/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/8491/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/3993
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/3993/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/3993/comments
|
https://api.github.com/repos/ollama/ollama/issues/3993/events
|
https://github.com/ollama/ollama/issues/3993
| 2,267,373,661
|
I_kwDOJ0Z1Ps6HJWBd
| 3,993
|
Add support for EMO-2B
|
{
"login": "OE-LUCIFER",
"id": 158988478,
"node_id": "U_kgDOCXn4vg",
"avatar_url": "https://avatars.githubusercontent.com/u/158988478?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/OE-LUCIFER",
"html_url": "https://github.com/OE-LUCIFER",
"followers_url": "https://api.github.com/users/OE-LUCIFER/followers",
"following_url": "https://api.github.com/users/OE-LUCIFER/following{/other_user}",
"gists_url": "https://api.github.com/users/OE-LUCIFER/gists{/gist_id}",
"starred_url": "https://api.github.com/users/OE-LUCIFER/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/OE-LUCIFER/subscriptions",
"organizations_url": "https://api.github.com/users/OE-LUCIFER/orgs",
"repos_url": "https://api.github.com/users/OE-LUCIFER/repos",
"events_url": "https://api.github.com/users/OE-LUCIFER/events{/privacy}",
"received_events_url": "https://api.github.com/users/OE-LUCIFER/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
open
| false
| null |
[] | null | 0
| 2024-04-28T06:21:00
| 2024-04-28T06:21:00
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Gguf model https://huggingface.co/Abhaykoul/EMO-2B-GGUF
Full model https://huggingface.co/OEvortex/EMO-2B
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/3993/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/3993/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/1376
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1376/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1376/comments
|
https://api.github.com/repos/ollama/ollama/issues/1376/events
|
https://github.com/ollama/ollama/pull/1376
| 2,024,549,360
|
PR_kwDOJ0Z1Ps5hF2eX
| 1,376
|
install: fix rocky kernel packages
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2023-12-04T19:20:27
| 2023-12-04T22:23:44
| 2023-12-04T22:23:43
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/1376",
"html_url": "https://github.com/ollama/ollama/pull/1376",
"diff_url": "https://github.com/ollama/ollama/pull/1376.diff",
"patch_url": "https://github.com/ollama/ollama/pull/1376.patch",
"merged_at": "2023-12-04T22:23:43"
}
|
package names for rocky-linux are slightly different
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1376/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1376/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/4537
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4537/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4537/comments
|
https://api.github.com/repos/ollama/ollama/issues/4537/events
|
https://github.com/ollama/ollama/issues/4537
| 2,305,927,857
|
I_kwDOJ0Z1Ps6Jcaqx
| 4,537
|
请问下如何将模型也封装进ollama的docker镜像中
|
{
"login": "iaoxuesheng",
"id": 94165844,
"node_id": "U_kgDOBZzbVA",
"avatar_url": "https://avatars.githubusercontent.com/u/94165844?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iaoxuesheng",
"html_url": "https://github.com/iaoxuesheng",
"followers_url": "https://api.github.com/users/iaoxuesheng/followers",
"following_url": "https://api.github.com/users/iaoxuesheng/following{/other_user}",
"gists_url": "https://api.github.com/users/iaoxuesheng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/iaoxuesheng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/iaoxuesheng/subscriptions",
"organizations_url": "https://api.github.com/users/iaoxuesheng/orgs",
"repos_url": "https://api.github.com/users/iaoxuesheng/repos",
"events_url": "https://api.github.com/users/iaoxuesheng/events{/privacy}",
"received_events_url": "https://api.github.com/users/iaoxuesheng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5789807732,
"node_id": "LA_kwDOJ0Z1Ps8AAAABWRl0dA",
"url": "https://api.github.com/repos/ollama/ollama/labels/model%20request",
"name": "model request",
"color": "1E5DE6",
"default": false,
"description": "Model requests"
}
] |
closed
| false
| null |
[] | null | 0
| 2024-05-20T12:48:17
| 2024-05-20T14:48:27
| 2024-05-20T14:48:27
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
请问下如何将qwen:32b模型也封装进ollama的docker镜像中
|
{
"login": "iaoxuesheng",
"id": 94165844,
"node_id": "U_kgDOBZzbVA",
"avatar_url": "https://avatars.githubusercontent.com/u/94165844?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/iaoxuesheng",
"html_url": "https://github.com/iaoxuesheng",
"followers_url": "https://api.github.com/users/iaoxuesheng/followers",
"following_url": "https://api.github.com/users/iaoxuesheng/following{/other_user}",
"gists_url": "https://api.github.com/users/iaoxuesheng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/iaoxuesheng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/iaoxuesheng/subscriptions",
"organizations_url": "https://api.github.com/users/iaoxuesheng/orgs",
"repos_url": "https://api.github.com/users/iaoxuesheng/repos",
"events_url": "https://api.github.com/users/iaoxuesheng/events{/privacy}",
"received_events_url": "https://api.github.com/users/iaoxuesheng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4537/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4537/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6042
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6042/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6042/comments
|
https://api.github.com/repos/ollama/ollama/issues/6042/events
|
https://github.com/ollama/ollama/issues/6042
| 2,434,948,575
|
I_kwDOJ0Z1Ps6RIl3f
| 6,042
|
strange tool response
|
{
"login": "asyncfncom",
"id": 136445484,
"node_id": "U_kgDOCCH-LA",
"avatar_url": "https://avatars.githubusercontent.com/u/136445484?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/asyncfncom",
"html_url": "https://github.com/asyncfncom",
"followers_url": "https://api.github.com/users/asyncfncom/followers",
"following_url": "https://api.github.com/users/asyncfncom/following{/other_user}",
"gists_url": "https://api.github.com/users/asyncfncom/gists{/gist_id}",
"starred_url": "https://api.github.com/users/asyncfncom/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/asyncfncom/subscriptions",
"organizations_url": "https://api.github.com/users/asyncfncom/orgs",
"repos_url": "https://api.github.com/users/asyncfncom/repos",
"events_url": "https://api.github.com/users/asyncfncom/events{/privacy}",
"received_events_url": "https://api.github.com/users/asyncfncom/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 2
| 2024-07-29T09:42:35
| 2024-08-15T21:44:53
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
The prompt was "call fibonacci function to determine 7 element of the sequence".
I wonder if there should be 2 tool calls.
```
{
"model": "llama3.1:8b",
"created_at": "2024-07-29T09:32:02.5425761Z",
"message": {
"role": "assistant",
"content": "",
"tool_calls": [
{
"function": {
"name": "fibonacci",
"arguments": {
"n": null
}
}
},
{
"function": {
"name": "fibonacci",
"arguments": {
"n": 7
}
}
}
]
},
"done_reason": "stop",
"done": true,
"total_duration": 17912684000,
"load_duration": 4394720300,
"prompt_eval_count": 132,
"prompt_eval_duration": 583579000,
"eval_count": 85,
"eval_duration": 12932451000
}
```
### OS
Windows
### GPU
Nvidia
### CPU
AMD
### Ollama version
0.3.0
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6042/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6042/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/7602
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/7602/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/7602/comments
|
https://api.github.com/repos/ollama/ollama/issues/7602/events
|
https://github.com/ollama/ollama/issues/7602
| 2,647,535,085
|
I_kwDOJ0Z1Ps6dzi3t
| 7,602
|
Ollama ps to report actual number of layers instead of percentage.
|
{
"login": "chigkim",
"id": 22120994,
"node_id": "MDQ6VXNlcjIyMTIwOTk0",
"avatar_url": "https://avatars.githubusercontent.com/u/22120994?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/chigkim",
"html_url": "https://github.com/chigkim",
"followers_url": "https://api.github.com/users/chigkim/followers",
"following_url": "https://api.github.com/users/chigkim/following{/other_user}",
"gists_url": "https://api.github.com/users/chigkim/gists{/gist_id}",
"starred_url": "https://api.github.com/users/chigkim/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/chigkim/subscriptions",
"organizations_url": "https://api.github.com/users/chigkim/orgs",
"repos_url": "https://api.github.com/users/chigkim/repos",
"events_url": "https://api.github.com/users/chigkim/events{/privacy}",
"received_events_url": "https://api.github.com/users/chigkim/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
open
| false
| null |
[] | null | 0
| 2024-11-10T19:02:27
| 2024-11-10T19:02:27
| null |
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Can we have Ollama to report how many layers out of total layers are offloaded to cpu instead of percentage?
I think This would be more useful information than just percentage when setting num_gpu parameter. Also you can see how many layers a model has.
Thanks!
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/7602/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/7602/timeline
| null | null | false
|
https://api.github.com/repos/ollama/ollama/issues/5617
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/5617/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/5617/comments
|
https://api.github.com/repos/ollama/ollama/issues/5617/events
|
https://github.com/ollama/ollama/pull/5617
| 2,401,858,365
|
PR_kwDOJ0Z1Ps51Bp_X
| 5,617
|
OpenAI: Update Docs to Include Tools
|
{
"login": "royjhan",
"id": 65097070,
"node_id": "MDQ6VXNlcjY1MDk3MDcw",
"avatar_url": "https://avatars.githubusercontent.com/u/65097070?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/royjhan",
"html_url": "https://github.com/royjhan",
"followers_url": "https://api.github.com/users/royjhan/followers",
"following_url": "https://api.github.com/users/royjhan/following{/other_user}",
"gists_url": "https://api.github.com/users/royjhan/gists{/gist_id}",
"starred_url": "https://api.github.com/users/royjhan/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/royjhan/subscriptions",
"organizations_url": "https://api.github.com/users/royjhan/orgs",
"repos_url": "https://api.github.com/users/royjhan/repos",
"events_url": "https://api.github.com/users/royjhan/events{/privacy}",
"received_events_url": "https://api.github.com/users/royjhan/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-07-10T22:39:15
| 2024-07-25T22:34:07
| 2024-07-25T22:34:06
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/5617",
"html_url": "https://github.com/ollama/ollama/pull/5617",
"diff_url": "https://github.com/ollama/ollama/pull/5617.diff",
"patch_url": "https://github.com/ollama/ollama/pull/5617.patch",
"merged_at": "2024-07-25T22:34:06"
}
| null |
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/5617/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/5617/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/2449
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2449/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2449/comments
|
https://api.github.com/repos/ollama/ollama/issues/2449/events
|
https://github.com/ollama/ollama/issues/2449
| 2,129,132,876
|
I_kwDOJ0Z1Ps5-5_1M
| 2,449
|
Log request/responses payload
|
{
"login": "jmformenti",
"id": 13070879,
"node_id": "MDQ6VXNlcjEzMDcwODc5",
"avatar_url": "https://avatars.githubusercontent.com/u/13070879?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmformenti",
"html_url": "https://github.com/jmformenti",
"followers_url": "https://api.github.com/users/jmformenti/followers",
"following_url": "https://api.github.com/users/jmformenti/following{/other_user}",
"gists_url": "https://api.github.com/users/jmformenti/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmformenti/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmformenti/subscriptions",
"organizations_url": "https://api.github.com/users/jmformenti/orgs",
"repos_url": "https://api.github.com/users/jmformenti/repos",
"events_url": "https://api.github.com/users/jmformenti/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmformenti/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 4
| 2024-02-11T19:26:43
| 2024-10-01T12:28:46
| 2024-05-11T00:36:44
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
In order to debug low-level details during development, it would be very useful to be able to see the payload of requests and responses.
Is there a way to enable this?
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2449/reactions",
"total_count": 1,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 1
}
|
https://api.github.com/repos/ollama/ollama/issues/2449/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2374
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2374/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2374/comments
|
https://api.github.com/repos/ollama/ollama/issues/2374/events
|
https://github.com/ollama/ollama/pull/2374
| 2,121,337,588
|
PR_kwDOJ0Z1Ps5mLF16
| 2,374
|
disable rocm builds
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-02-06T17:29:51
| 2024-02-06T17:41:04
| 2024-02-06T17:41:03
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/2374",
"html_url": "https://github.com/ollama/ollama/pull/2374",
"diff_url": "https://github.com/ollama/ollama/pull/2374.diff",
"patch_url": "https://github.com/ollama/ollama/pull/2374.patch",
"merged_at": "2024-02-06T17:41:03"
}
|
rocm builds are failing because of disk space issues. disable them temporarily until larger runners
resolves #2373
|
{
"login": "mxyng",
"id": 2372640,
"node_id": "MDQ6VXNlcjIzNzI2NDA=",
"avatar_url": "https://avatars.githubusercontent.com/u/2372640?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/mxyng",
"html_url": "https://github.com/mxyng",
"followers_url": "https://api.github.com/users/mxyng/followers",
"following_url": "https://api.github.com/users/mxyng/following{/other_user}",
"gists_url": "https://api.github.com/users/mxyng/gists{/gist_id}",
"starred_url": "https://api.github.com/users/mxyng/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/mxyng/subscriptions",
"organizations_url": "https://api.github.com/users/mxyng/orgs",
"repos_url": "https://api.github.com/users/mxyng/repos",
"events_url": "https://api.github.com/users/mxyng/events{/privacy}",
"received_events_url": "https://api.github.com/users/mxyng/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2374/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2374/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1458
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1458/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1458/comments
|
https://api.github.com/repos/ollama/ollama/issues/1458/events
|
https://github.com/ollama/ollama/issues/1458
| 2,034,704,639
|
I_kwDOJ0Z1Ps55RyD_
| 1,458
|
Ollama hung after 30 minute of use
|
{
"login": "lfoppiano",
"id": 15426,
"node_id": "MDQ6VXNlcjE1NDI2",
"avatar_url": "https://avatars.githubusercontent.com/u/15426?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/lfoppiano",
"html_url": "https://github.com/lfoppiano",
"followers_url": "https://api.github.com/users/lfoppiano/followers",
"following_url": "https://api.github.com/users/lfoppiano/following{/other_user}",
"gists_url": "https://api.github.com/users/lfoppiano/gists{/gist_id}",
"starred_url": "https://api.github.com/users/lfoppiano/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/lfoppiano/subscriptions",
"organizations_url": "https://api.github.com/users/lfoppiano/orgs",
"repos_url": "https://api.github.com/users/lfoppiano/repos",
"events_url": "https://api.github.com/users/lfoppiano/events{/privacy}",
"received_events_url": "https://api.github.com/users/lfoppiano/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 22
| 2023-12-11T02:28:13
| 2024-05-05T01:11:36
| 2024-02-20T01:20:04
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
I'm running Ollama on my mac M1 and I'm trying to use the 7b models for processing batches of questions / answers.
I noticed that after a while ollama just hang and the process stay there forever.
Is there a way to know what's going on?
I did not find a way to get to the logs.
Thank you in advance
|
{
"login": "jmorganca",
"id": 251292,
"node_id": "MDQ6VXNlcjI1MTI5Mg==",
"avatar_url": "https://avatars.githubusercontent.com/u/251292?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/jmorganca",
"html_url": "https://github.com/jmorganca",
"followers_url": "https://api.github.com/users/jmorganca/followers",
"following_url": "https://api.github.com/users/jmorganca/following{/other_user}",
"gists_url": "https://api.github.com/users/jmorganca/gists{/gist_id}",
"starred_url": "https://api.github.com/users/jmorganca/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/jmorganca/subscriptions",
"organizations_url": "https://api.github.com/users/jmorganca/orgs",
"repos_url": "https://api.github.com/users/jmorganca/repos",
"events_url": "https://api.github.com/users/jmorganca/events{/privacy}",
"received_events_url": "https://api.github.com/users/jmorganca/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1458/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1458/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4255
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4255/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4255/comments
|
https://api.github.com/repos/ollama/ollama/issues/4255/events
|
https://github.com/ollama/ollama/issues/4255
| 2,285,208,246
|
I_kwDOJ0Z1Ps6INYK2
| 4,255
|
max retries exceeded: http status 502 Bad Gateway while pushing a model
|
{
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/taozhiyuai/followers",
"following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}",
"gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}",
"starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions",
"organizations_url": "https://api.github.com/users/taozhiyuai/orgs",
"repos_url": "https://api.github.com/users/taozhiyuai/repos",
"events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}",
"received_events_url": "https://api.github.com/users/taozhiyuai/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
closed
| false
| null |
[] | null | 0
| 2024-05-08T10:05:11
| 2024-05-10T12:17:36
| 2024-05-10T12:17:36
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
I have spend nearly a whole day to push, but fail. are there anyway to set the time of re-try? please cancel limitation of re-tries. or is that possible pushing is resumable?
taozhiyu@603e5f4a42f1 Q8 % ollama push taozhiyuai/openbiollm-llama-3-70b:q8_0
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
retrieving manifest
pushing cf7fb8dac551... 13% ▕█████ ▏ 10 GB/ 74 GB 4.3 MB/s 4h13m
Error: max retries exceeded: http status 502 Bad Gateway: <html>
<head><title>502 Bad Gateway</title></head>
<body>
<center><h1>502 Bad Gateway</h1></center>
<hr><center>cloudflare</center>
</body>
</html>
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
up to date
|
{
"login": "taozhiyuai",
"id": 146583103,
"node_id": "U_kgDOCLyuPw",
"avatar_url": "https://avatars.githubusercontent.com/u/146583103?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/taozhiyuai",
"html_url": "https://github.com/taozhiyuai",
"followers_url": "https://api.github.com/users/taozhiyuai/followers",
"following_url": "https://api.github.com/users/taozhiyuai/following{/other_user}",
"gists_url": "https://api.github.com/users/taozhiyuai/gists{/gist_id}",
"starred_url": "https://api.github.com/users/taozhiyuai/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/taozhiyuai/subscriptions",
"organizations_url": "https://api.github.com/users/taozhiyuai/orgs",
"repos_url": "https://api.github.com/users/taozhiyuai/repos",
"events_url": "https://api.github.com/users/taozhiyuai/events{/privacy}",
"received_events_url": "https://api.github.com/users/taozhiyuai/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4255/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4255/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2022
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2022/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2022/comments
|
https://api.github.com/repos/ollama/ollama/issues/2022/events
|
https://github.com/ollama/ollama/issues/2022
| 2,084,797,403
|
I_kwDOJ0Z1Ps58Q3vb
| 2,022
|
List available models
|
{
"login": "ParisNeo",
"id": 827993,
"node_id": "MDQ6VXNlcjgyNzk5Mw==",
"avatar_url": "https://avatars.githubusercontent.com/u/827993?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/ParisNeo",
"html_url": "https://github.com/ParisNeo",
"followers_url": "https://api.github.com/users/ParisNeo/followers",
"following_url": "https://api.github.com/users/ParisNeo/following{/other_user}",
"gists_url": "https://api.github.com/users/ParisNeo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/ParisNeo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/ParisNeo/subscriptions",
"organizations_url": "https://api.github.com/users/ParisNeo/orgs",
"repos_url": "https://api.github.com/users/ParisNeo/repos",
"events_url": "https://api.github.com/users/ParisNeo/events{/privacy}",
"received_events_url": "https://api.github.com/users/ParisNeo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
},
{
"id": 6573197867,
"node_id": "LA_kwDOJ0Z1Ps8AAAABh8sKKw",
"url": "https://api.github.com/repos/ollama/ollama/labels/ollama.com",
"name": "ollama.com",
"color": "ffffff",
"default": false,
"description": ""
}
] |
closed
| false
| null |
[] | null | 16
| 2024-01-16T20:14:24
| 2024-11-21T17:26:22
| 2024-11-21T17:26:22
|
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Hi. The API allows me to list the local models. Is there a way to list all available models (those we can find in the website of ollama?
I need that for the models zoo to make it easy for users of lollms with ollama backend to install the models.
I prefer this rather than having to scrape the website to get the latest list of models.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2022/reactions",
"total_count": 23,
"+1": 23,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2022/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6284
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6284/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6284/comments
|
https://api.github.com/repos/ollama/ollama/issues/6284/events
|
https://github.com/ollama/ollama/issues/6284
| 2,457,788,664
|
I_kwDOJ0Z1Ps6SfuD4
| 6,284
|
Intel GPU in Docker container crashes
|
{
"login": "Minionflo",
"id": 62773986,
"node_id": "MDQ6VXNlcjYyNzczOTg2",
"avatar_url": "https://avatars.githubusercontent.com/u/62773986?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Minionflo",
"html_url": "https://github.com/Minionflo",
"followers_url": "https://api.github.com/users/Minionflo/followers",
"following_url": "https://api.github.com/users/Minionflo/following{/other_user}",
"gists_url": "https://api.github.com/users/Minionflo/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Minionflo/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Minionflo/subscriptions",
"organizations_url": "https://api.github.com/users/Minionflo/orgs",
"repos_url": "https://api.github.com/users/Minionflo/repos",
"events_url": "https://api.github.com/users/Minionflo/events{/privacy}",
"received_events_url": "https://api.github.com/users/Minionflo/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
},
{
"id": 6677491450,
"node_id": "LA_kwDOJ0Z1Ps8AAAABjgJu-g",
"url": "https://api.github.com/repos/ollama/ollama/labels/intel",
"name": "intel",
"color": "226E5B",
"default": false,
"description": "issues relating to Intel GPUs"
}
] |
closed
| false
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
] | null | 1
| 2024-08-09T12:14:56
| 2024-08-09T19:14:44
| 2024-08-09T19:14:44
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
### What is the issue?
with the error `panic: runtime error: invalid memory address or nil pointer dereference`
Docker Compose File: https://bin.minionflo.net/p/E9gFhE.yaml
Log: https://bin.minionflo.net/p/QyrT8Z.txt
### OS
Docker on Linux
### GPU
Intel
### CPU
AMD
### Ollama version
0.3.4
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6284/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6284/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/6432
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/6432/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/6432/comments
|
https://api.github.com/repos/ollama/ollama/issues/6432/events
|
https://github.com/ollama/ollama/pull/6432
| 2,474,477,204
|
PR_kwDOJ0Z1Ps54y56d
| 6,432
|
Split rocm back out of bundle
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 0
| 2024-08-20T00:12:54
| 2024-08-20T14:26:41
| 2024-08-20T14:26:38
|
COLLABORATOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | false
|
{
"url": "https://api.github.com/repos/ollama/ollama/pulls/6432",
"html_url": "https://github.com/ollama/ollama/pull/6432",
"diff_url": "https://github.com/ollama/ollama/pull/6432.diff",
"patch_url": "https://github.com/ollama/ollama/pull/6432.patch",
"merged_at": "2024-08-20T14:26:38"
}
|
We're [over budget for github's maximum release artifact size](https://github.com/ollama/ollama/actions/runs/10461795539/job/28973022210) with rocm + 2 cuda versions. This splits rocm back out as a discrete artifact, but keeps the layout so it can be extracted into the same location as the main bundle.
|
{
"login": "dhiltgen",
"id": 4033016,
"node_id": "MDQ6VXNlcjQwMzMwMTY=",
"avatar_url": "https://avatars.githubusercontent.com/u/4033016?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/dhiltgen",
"html_url": "https://github.com/dhiltgen",
"followers_url": "https://api.github.com/users/dhiltgen/followers",
"following_url": "https://api.github.com/users/dhiltgen/following{/other_user}",
"gists_url": "https://api.github.com/users/dhiltgen/gists{/gist_id}",
"starred_url": "https://api.github.com/users/dhiltgen/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/dhiltgen/subscriptions",
"organizations_url": "https://api.github.com/users/dhiltgen/orgs",
"repos_url": "https://api.github.com/users/dhiltgen/repos",
"events_url": "https://api.github.com/users/dhiltgen/events{/privacy}",
"received_events_url": "https://api.github.com/users/dhiltgen/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/6432/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/6432/timeline
| null | null | true
|
https://api.github.com/repos/ollama/ollama/issues/1346
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/1346/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/1346/comments
|
https://api.github.com/repos/ollama/ollama/issues/1346/events
|
https://github.com/ollama/ollama/issues/1346
| 2,021,266,981
|
I_kwDOJ0Z1Ps54ehYl
| 1,346
|
Set conversation or chat history/context in CLI
|
{
"login": "Maharshi-Pandya",
"id": 53078775,
"node_id": "MDQ6VXNlcjUzMDc4Nzc1",
"avatar_url": "https://avatars.githubusercontent.com/u/53078775?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Maharshi-Pandya",
"html_url": "https://github.com/Maharshi-Pandya",
"followers_url": "https://api.github.com/users/Maharshi-Pandya/followers",
"following_url": "https://api.github.com/users/Maharshi-Pandya/following{/other_user}",
"gists_url": "https://api.github.com/users/Maharshi-Pandya/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Maharshi-Pandya/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Maharshi-Pandya/subscriptions",
"organizations_url": "https://api.github.com/users/Maharshi-Pandya/orgs",
"repos_url": "https://api.github.com/users/Maharshi-Pandya/repos",
"events_url": "https://api.github.com/users/Maharshi-Pandya/events{/privacy}",
"received_events_url": "https://api.github.com/users/Maharshi-Pandya/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 1
| 2023-12-01T17:03:50
| 2023-12-27T15:09:54
| 2023-12-27T15:09:54
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Thank you for making this!
I tried the `/set history` command within the CLI and expected it to work.
I would like to use the CLI as a chatbot itself having access to conversation history (a window of messages if not whole).
What is the process to set the conversation history as context in `Openhermes-mistral` specifically?
Here is the bug:
<img width="1070" alt="image" src="https://github.com/jmorganca/ollama/assets/53078775/c6033c8b-5054-41a8-9977-9a0e0ab8b130">
|
{
"login": "Maharshi-Pandya",
"id": 53078775,
"node_id": "MDQ6VXNlcjUzMDc4Nzc1",
"avatar_url": "https://avatars.githubusercontent.com/u/53078775?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/Maharshi-Pandya",
"html_url": "https://github.com/Maharshi-Pandya",
"followers_url": "https://api.github.com/users/Maharshi-Pandya/followers",
"following_url": "https://api.github.com/users/Maharshi-Pandya/following{/other_user}",
"gists_url": "https://api.github.com/users/Maharshi-Pandya/gists{/gist_id}",
"starred_url": "https://api.github.com/users/Maharshi-Pandya/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/Maharshi-Pandya/subscriptions",
"organizations_url": "https://api.github.com/users/Maharshi-Pandya/orgs",
"repos_url": "https://api.github.com/users/Maharshi-Pandya/repos",
"events_url": "https://api.github.com/users/Maharshi-Pandya/events{/privacy}",
"received_events_url": "https://api.github.com/users/Maharshi-Pandya/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/1346/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/1346/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/4458
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/4458/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/4458/comments
|
https://api.github.com/repos/ollama/ollama/issues/4458/events
|
https://github.com/ollama/ollama/issues/4458
| 2,298,859,750
|
I_kwDOJ0Z1Ps6JBdDm
| 4,458
|
Confirm GPU usage command
|
{
"login": "puddlejumper90",
"id": 55165215,
"node_id": "MDQ6VXNlcjU1MTY1MjE1",
"avatar_url": "https://avatars.githubusercontent.com/u/55165215?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/puddlejumper90",
"html_url": "https://github.com/puddlejumper90",
"followers_url": "https://api.github.com/users/puddlejumper90/followers",
"following_url": "https://api.github.com/users/puddlejumper90/following{/other_user}",
"gists_url": "https://api.github.com/users/puddlejumper90/gists{/gist_id}",
"starred_url": "https://api.github.com/users/puddlejumper90/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/puddlejumper90/subscriptions",
"organizations_url": "https://api.github.com/users/puddlejumper90/orgs",
"repos_url": "https://api.github.com/users/puddlejumper90/repos",
"events_url": "https://api.github.com/users/puddlejumper90/events{/privacy}",
"received_events_url": "https://api.github.com/users/puddlejumper90/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396200,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aaA",
"url": "https://api.github.com/repos/ollama/ollama/labels/feature%20request",
"name": "feature request",
"color": "a2eeef",
"default": false,
"description": "New feature or request"
}
] |
closed
| false
| null |
[] | null | 3
| 2024-05-15T21:16:25
| 2024-05-16T21:11:45
| 2024-05-15T22:53:09
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
Would like to have a way to to confirm if a GPU is actually being utilized. Maybe some kind of command or option when running a given model to test/log individual machine performance.
|
{
"login": "pdevine",
"id": 75239,
"node_id": "MDQ6VXNlcjc1MjM5",
"avatar_url": "https://avatars.githubusercontent.com/u/75239?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/pdevine",
"html_url": "https://github.com/pdevine",
"followers_url": "https://api.github.com/users/pdevine/followers",
"following_url": "https://api.github.com/users/pdevine/following{/other_user}",
"gists_url": "https://api.github.com/users/pdevine/gists{/gist_id}",
"starred_url": "https://api.github.com/users/pdevine/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/pdevine/subscriptions",
"organizations_url": "https://api.github.com/users/pdevine/orgs",
"repos_url": "https://api.github.com/users/pdevine/repos",
"events_url": "https://api.github.com/users/pdevine/events{/privacy}",
"received_events_url": "https://api.github.com/users/pdevine/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/4458/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/4458/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2366
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2366/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2366/comments
|
https://api.github.com/repos/ollama/ollama/issues/2366/events
|
https://github.com/ollama/ollama/issues/2366
| 2,119,725,248
|
I_kwDOJ0Z1Ps5-WHDA
| 2,366
|
Bump llama.cpp commit to 6b91b1e which includes Intel GPU support (iGPU, Arc, Max, Flex)
|
{
"login": "0x33taji",
"id": 148982823,
"node_id": "U_kgDOCOFMJw",
"avatar_url": "https://avatars.githubusercontent.com/u/148982823?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/0x33taji",
"html_url": "https://github.com/0x33taji",
"followers_url": "https://api.github.com/users/0x33taji/followers",
"following_url": "https://api.github.com/users/0x33taji/following{/other_user}",
"gists_url": "https://api.github.com/users/0x33taji/gists{/gist_id}",
"starred_url": "https://api.github.com/users/0x33taji/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/0x33taji/subscriptions",
"organizations_url": "https://api.github.com/users/0x33taji/orgs",
"repos_url": "https://api.github.com/users/0x33taji/repos",
"events_url": "https://api.github.com/users/0x33taji/events{/privacy}",
"received_events_url": "https://api.github.com/users/0x33taji/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[] |
closed
| false
| null |
[] | null | 2
| 2024-02-06T00:44:53
| 2024-02-13T21:52:10
| 2024-02-13T21:52:10
|
NONE
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
llama.cpp has added support for Intel GPUs.
commit ID: [6b91b1e0a92ac2e4e269eec6361ca53a61ced6c6](https://github.com/ggerganov/llama.cpp/commit/6b91b1e0a92ac2e4e269eec6361ca53a61ced6c6)
*Task*
1. Bump llama.cpp commit if feasible
2. Then update Dockerfile with with Intel GPU support for one-click deployment or as reference to bare metal deployment.
**Reference for dockerfile implementation**
llama.cpp guidelines on Intel GPU support via SYCL lib.
- https://github.com/ggerganov/llama.cpp/blob/master/README-sycl.md
|
{
"login": "0x33taji",
"id": 148982823,
"node_id": "U_kgDOCOFMJw",
"avatar_url": "https://avatars.githubusercontent.com/u/148982823?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/0x33taji",
"html_url": "https://github.com/0x33taji",
"followers_url": "https://api.github.com/users/0x33taji/followers",
"following_url": "https://api.github.com/users/0x33taji/following{/other_user}",
"gists_url": "https://api.github.com/users/0x33taji/gists{/gist_id}",
"starred_url": "https://api.github.com/users/0x33taji/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/0x33taji/subscriptions",
"organizations_url": "https://api.github.com/users/0x33taji/orgs",
"repos_url": "https://api.github.com/users/0x33taji/repos",
"events_url": "https://api.github.com/users/0x33taji/events{/privacy}",
"received_events_url": "https://api.github.com/users/0x33taji/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2366/reactions",
"total_count": 0,
"+1": 0,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2366/timeline
| null |
completed
| false
|
https://api.github.com/repos/ollama/ollama/issues/2653
|
https://api.github.com/repos/ollama/ollama
|
https://api.github.com/repos/ollama/ollama/issues/2653/labels{/name}
|
https://api.github.com/repos/ollama/ollama/issues/2653/comments
|
https://api.github.com/repos/ollama/ollama/issues/2653/events
|
https://github.com/ollama/ollama/issues/2653
| 2,147,711,815
|
I_kwDOJ0Z1Ps6AA3tH
| 2,653
|
Ollama serve fails silently when an input is too long
|
{
"login": "logancyang",
"id": 4860545,
"node_id": "MDQ6VXNlcjQ4NjA1NDU=",
"avatar_url": "https://avatars.githubusercontent.com/u/4860545?v=4",
"gravatar_id": "",
"url": "https://api.github.com/users/logancyang",
"html_url": "https://github.com/logancyang",
"followers_url": "https://api.github.com/users/logancyang/followers",
"following_url": "https://api.github.com/users/logancyang/following{/other_user}",
"gists_url": "https://api.github.com/users/logancyang/gists{/gist_id}",
"starred_url": "https://api.github.com/users/logancyang/starred{/owner}{/repo}",
"subscriptions_url": "https://api.github.com/users/logancyang/subscriptions",
"organizations_url": "https://api.github.com/users/logancyang/orgs",
"repos_url": "https://api.github.com/users/logancyang/repos",
"events_url": "https://api.github.com/users/logancyang/events{/privacy}",
"received_events_url": "https://api.github.com/users/logancyang/received_events",
"type": "User",
"user_view_type": "public",
"site_admin": false
}
|
[
{
"id": 5667396184,
"node_id": "LA_kwDOJ0Z1Ps8AAAABUc2aWA",
"url": "https://api.github.com/repos/ollama/ollama/labels/bug",
"name": "bug",
"color": "d73a4a",
"default": true,
"description": "Something isn't working"
}
] |
open
| false
| null |
[] | null | 6
| 2024-02-21T21:05:18
| 2024-03-12T02:02:12
| null |
CONTRIBUTOR
|
{
"total": 0,
"completed": 0,
"percent_completed": 0
}
| null | null | null |
When I use `ollama serve` and provide a context of ~30k tokens with a mistral model that has a max context window of 32768, the server doesn't show any error and proceeds to return as usual. That gave me the impression that it successfully took in the entire context.
But after digging a bit deeper, I see it's not.

So when I do this below it started working fine
```
ollama run <model>
/set parameter num_ctx 32768
/save
```
Perhaps it's because there are flags to set with `ollama serve` which I don't know about after reading the docs. Is there a better way to set the context window for `ollama serve`?
In my mind, the expected behavior is to show an error message when the input is exceeding the set context window length. LM Studio does this
<img width="1426" alt="SCR-20240221-lsnn" src="https://github.com/ollama/ollama/assets/4860545/ee4f2408-bbce-4fb8-bd74-6306aca08b3c">
Please let me know if it's because I'm not using it with the right flags or if this is a legit concern.
| null |
{
"url": "https://api.github.com/repos/ollama/ollama/issues/2653/reactions",
"total_count": 2,
"+1": 2,
"-1": 0,
"laugh": 0,
"hooray": 0,
"confused": 0,
"heart": 0,
"rocket": 0,
"eyes": 0
}
|
https://api.github.com/repos/ollama/ollama/issues/2653/timeline
| null | null | false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.