choco-conoz commited on
Commit
d2a22fe
·
1 Parent(s): e117631

feat: change model

Browse files
Files changed (2) hide show
  1. Dockerfile +1 -15
  2. src/streamlit_app.py +2 -42
Dockerfile CHANGED
@@ -10,16 +10,11 @@ RUN apt-get update && apt-get install -y \
10
  git \
11
  && rm -rf /var/lib/apt/lists/*
12
 
13
- # ENV PYTHONUNBUFFERED=1 \
14
- # PORT=8000 \
15
- # HF_HOME=/home/user/huggingface
16
-
17
  COPY src/ ./src/
18
- # RUN pip3 install poetry==2.1.3
19
- # RUN pip install -U "huggingface_hub"
20
  COPY requirements.txt ./
21
  RUN pip3 install --no-cache-dir --upgrade -r requirements.txt
22
 
 
23
  RUN useradd -m -u 1000 user
24
  USER user
25
  ENV HOME=/home/user \
@@ -27,15 +22,6 @@ ENV HOME=/home/user \
27
  WORKDIR $HOME/app
28
  COPY --chown=user . $HOME/app
29
 
30
- # COPY pyproject.toml README.md /app
31
- # RUN poetry install
32
- # RUN poetry config virtualenvs.in-project true
33
-
34
- # ENV VIRTUAL_ENV=/app/.venv
35
- # ENV PATH="$VIRTUAL_ENV/bin:$PATH"
36
-
37
  EXPOSE 8501
38
-
39
  HEALTHCHECK CMD curl --fail http://localhost:8501/_stcore/health
40
-
41
  ENTRYPOINT ["streamlit", "run", "src/streamlit_app.py", "--server.port=8501", "--server.address=0.0.0.0"]
 
10
  git \
11
  && rm -rf /var/lib/apt/lists/*
12
 
 
 
 
 
13
  COPY src/ ./src/
 
 
14
  COPY requirements.txt ./
15
  RUN pip3 install --no-cache-dir --upgrade -r requirements.txt
16
 
17
+ # for cache permission
18
  RUN useradd -m -u 1000 user
19
  USER user
20
  ENV HOME=/home/user \
 
22
  WORKDIR $HOME/app
23
  COPY --chown=user . $HOME/app
24
 
 
 
 
 
 
 
 
25
  EXPOSE 8501
 
26
  HEALTHCHECK CMD curl --fail http://localhost:8501/_stcore/health
 
27
  ENTRYPOINT ["streamlit", "run", "src/streamlit_app.py", "--server.port=8501", "--server.address=0.0.0.0"]
src/streamlit_app.py CHANGED
@@ -1,26 +1,12 @@
1
- import os
2
  import streamlit as st
3
  import torch
4
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
5
- # from huggingface_hub import notebook_login
6
- # from unsloth import FastLanguageModel, is_bfloat16_supported
7
 
8
  # model_id = "sentence-transformers/all-MiniLM-L6-v2"
9
  # model_id = "sentence-transformers/xlm-r-base-en-ko-nli-ststb"
10
-
11
  # model_id = "mistralai/Mistral-7B-Instruct-v0.1"
12
- # model_id = "meta-llama/Llama-3.2-1B"
13
- model_id = "choco-conoz/TwinLlama-3.1-8B"
14
-
15
- # processor = pipeline(
16
- # "text-generation",
17
- # model=model_id,
18
- # model_kwargs={
19
- # "torch_dtype": torch.float16,
20
- # "quantization_config": {"load_in_4bit": True},
21
- # "low_cpu_mem_usage": True,
22
- # },
23
- # )
24
 
25
  tokenizer = AutoTokenizer.from_pretrained(model_id)
26
  model = AutoModelForCausalLM.from_pretrained(model_id)
@@ -67,29 +53,3 @@ def main():
67
 
68
  if __name__ == "__main__":
69
  main()
70
- # >>> old
71
- # num_points = st.slider("Number of points in spiral", 1, 10000, 1100)
72
- # num_turns = st.slider("Number of turns in spiral", 1, 300, 31)
73
-
74
- # indices = np.linspace(0, 1, num_points)
75
- # theta = 2 * np.pi * num_turns * indices
76
- # radius = indices
77
-
78
- # x = radius * np.cos(theta)
79
- # y = radius * np.sin(theta)
80
-
81
- # df = pd.DataFrame({
82
- # "x": x,
83
- # "y": y,
84
- # "idx": indices,
85
- # "rand": np.random.randn(num_points),
86
- # })
87
-
88
- # st.altair_chart(alt.Chart(df, height=700, width=700)
89
- # .mark_point(filled=True)
90
- # .encode(
91
- # x=alt.X("x", axis=None),
92
- # y=alt.Y("y", axis=None),
93
- # color=alt.Color("idx", legend=None, scale=alt.Scale()),
94
- # size=alt.Size("rand", legend=None, scale=alt.Scale(range=[1, 150])),
95
- # ))
 
 
1
  import streamlit as st
2
  import torch
3
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
 
 
4
 
5
  # model_id = "sentence-transformers/all-MiniLM-L6-v2"
6
  # model_id = "sentence-transformers/xlm-r-base-en-ko-nli-ststb"
 
7
  # model_id = "mistralai/Mistral-7B-Instruct-v0.1"
8
+ model_id = "meta-llama/Llama-3.2-1B"
9
+ # model_id = "choco-conoz/TwinLlama-3.1-8B"
 
 
 
 
 
 
 
 
 
 
10
 
11
  tokenizer = AutoTokenizer.from_pretrained(model_id)
12
  model = AutoModelForCausalLM.from_pretrained(model_id)
 
53
 
54
  if __name__ == "__main__":
55
  main()