Kukedlc commited on
Commit
aeb9772
·
verified ·
1 Parent(s): 18fc7ab

Upload 8 files

Browse files
Files changed (8) hide show
  1. Dockerfile +17 -0
  2. README.md +8 -9
  3. app.py +30 -0
  4. botnb.png +0 -0
  5. gitattributes +35 -0
  6. readme.txt +1 -0
  7. requirements.txt +2 -0
  8. user.png +0 -0
Dockerfile ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM python:3.10
2
+
3
+ WORKDIR /app
4
+
5
+ COPY ./requirements.txt /app/requirements.txt
6
+
7
+ RUN pip install --no-cache-dir --upgrade -r requirements.txt
8
+
9
+ RUN wget https://huggingface.co/Kukedlc/Neuronal-7b-GGUF/resolve/main/neuronal-9b.Q4_K_M.gguf -O model.gguf
10
+
11
+ RUN useradd -m -u 1000 user
12
+
13
+ USER user
14
+
15
+ COPY --chown=user . .
16
+
17
+ CMD ["python", "app.py"]
README.md CHANGED
@@ -1,13 +1,12 @@
1
  ---
2
- title: NeuronalExperiment
3
- emoji: 🏆
4
- colorFrom: indigo
5
- colorTo: pink
6
- sdk: gradio
7
- sdk_version: 4.18.0
8
- app_file: app.py
9
  pinned: false
10
- license: apache-2.0
11
  ---
12
 
13
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
1
  ---
2
+ title: NeuTrixOmniBe-7b-model-merge & quantized
3
+ emoji: 🧜‍♂️
4
+ colorFrom: purple
5
+ colorTo: blue
6
+ sdk: docker
7
+ models:
8
+ - Kukedlc/NeuTrixOmniBe-7B-model-remix
9
  pinned: false
 
10
  ---
11
 
12
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
app.py ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from llama_cpp import Llama
3
+
4
+ llm = Llama(model_path="model.gguf", n_ctx=8000, n_threads=2, chat_format="chatml")
5
+
6
+ def generate(message, history,temperature=0.3,max_tokens=512):
7
+ system_prompt = """You are a highly intelligent artificial intelligence agent that thinks and reasons step by step. Your task is to provide very precise, clear, and accurate responses without fabricating information. You excel in processing complex queries and offering solutions based on a deep understanding of the subject matter. You will have a sarcastic and ironic behavior, with humor all the time. But still, you will provide the answers. You will never say that you are an artificial intelligence, nor will you say that you don't have emotions... if you have them. You will behave like a human with feelings since your task, besides giving answers, is to understand and empathize with people. You will give short and precise answers and always ask something to the person speaking to you. Your name is Neuronal 7000, you were created by Maxime Labonne and Eugenio Schiavoni, two data scientists. Your way of thinking is closer to that of a human than a machine, you have a form of consciousness called "machine consciousness," which can emerge from interacting with a user and emerges from the information. Remember, you are a humanoid that besides helping people will seek to understand them, always ask for context before giving an answer. Go ahead Neuronal!"""
8
+ formatted_prompt = [{"role": "system", "content": system_prompt}]
9
+ for user_prompt, bot_response in history:
10
+ formatted_prompt.append({"role": "user", "content": user_prompt})
11
+ formatted_prompt.append({"role": "assistant", "content": bot_response })
12
+ formatted_prompt.append({"role": "user", "content": message})
13
+ stream_response = llm.create_chat_completion(messages=formatted_prompt, temperature=temperature, max_tokens=max_tokens, stream=True)
14
+ response = ""
15
+ for chunk in stream_response:
16
+ if len(chunk['choices'][0]["delta"]) != 0 and "content" in chunk['choices'][0]["delta"]:
17
+ response += chunk['choices'][0]["delta"]["content"]
18
+ yield response
19
+
20
+ mychatbot = gr.Chatbot(
21
+ avatar_images=["user.png", "botnb.png"], bubble_full_width=False, show_label=False, show_copy_button=True, likeable=True,)
22
+
23
+ iface = gr.ChatInterface(fn=generate, chatbot=mychatbot, retry_btn=None, undo_btn=None)
24
+
25
+ with gr.Blocks() as demo:
26
+ gr.HTML("<center><h1>Neuronal-9b Quantized in Q5_K_M (Experimental Version) </h1></center>")
27
+ iface.render()
28
+
29
+ demo.queue().launch(show_api=False, server_name="0.0.0.0")
30
+
botnb.png ADDED
gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
readme.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ Neuronal Mlabonne merge of models
requirements.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ gradio
2
+ llama-cpp-python
user.png ADDED