rewgwrth commited on
Commit
2c01a1c
·
1 Parent(s): 6e256a4
.idea/.gitignore ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ # Default ignored files
2
+ /shelf/
3
+ /workspace.xml
.idea/Game-Engine-3-LLM.iml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0" encoding="UTF-8"?>
2
+ <module type="PYTHON_MODULE" version="4">
3
+ <component name="NewModuleRootManager">
4
+ <content url="file://$MODULE_DIR$">
5
+ <excludeFolder url="file://$MODULE_DIR$/.venv" />
6
+ </content>
7
+ <orderEntry type="inheritedJdk" />
8
+ <orderEntry type="sourceFolder" forTests="false" />
9
+ </component>
10
+ </module>
.idea/inspectionProfiles/Project_Default.xml ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <component name="InspectionProjectProfileManager">
2
+ <profile version="1.0">
3
+ <option name="myName" value="Project Default" />
4
+ <inspection_tool class="GrazieInspection" enabled="false" level="GRAMMAR_ERROR" enabled_by_default="false" />
5
+ <inspection_tool class="PyBroadExceptionInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false" />
6
+ <inspection_tool class="PyDictCreationInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false" />
7
+ <inspection_tool class="PyInterpreterInspection" enabled="true" level="INFORMATION" enabled_by_default="true" />
8
+ <inspection_tool class="PyMethodMayBeStaticInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false" />
9
+ <inspection_tool class="PyPackageRequirementsInspection" enabled="true" level="WARNING" enabled_by_default="true">
10
+ <option name="ignoredPackages">
11
+ <value>
12
+ <list size="20">
13
+ <item index="0" class="java.lang.String" itemvalue="setuptools" />
14
+ <item index="1" class="java.lang.String" itemvalue="pillow" />
15
+ <item index="2" class="java.lang.String" itemvalue="python-dotenv" />
16
+ <item index="3" class="java.lang.String" itemvalue="Flask" />
17
+ <item index="4" class="java.lang.String" itemvalue="pywin32" />
18
+ <item index="5" class="java.lang.String" itemvalue="pyinstaller" />
19
+ <item index="6" class="java.lang.String" itemvalue="numba" />
20
+ <item index="7" class="java.lang.String" itemvalue="pyperclip" />
21
+ <item index="8" class="java.lang.String" itemvalue="cython" />
22
+ <item index="9" class="java.lang.String" itemvalue="orjson" />
23
+ <item index="10" class="java.lang.String" itemvalue="llvmlite" />
24
+ <item index="11" class="java.lang.String" itemvalue="numpy" />
25
+ <item index="12" class="java.lang.String" itemvalue="requests" />
26
+ <item index="13" class="java.lang.String" itemvalue="pynput" />
27
+ <item index="14" class="java.lang.String" itemvalue="certifi" />
28
+ <item index="15" class="java.lang.String" itemvalue="PyQt5_sip" />
29
+ <item index="16" class="java.lang.String" itemvalue="charset-normalizer" />
30
+ <item index="17" class="java.lang.String" itemvalue="urllib3" />
31
+ <item index="18" class="java.lang.String" itemvalue="idna" />
32
+ <item index="19" class="java.lang.String" itemvalue="ujson" />
33
+ </list>
34
+ </value>
35
+ </option>
36
+ </inspection_tool>
37
+ <inspection_tool class="PyPep8NamingInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false">
38
+ <option name="ignoredErrors">
39
+ <list>
40
+ <option value="N803" />
41
+ </list>
42
+ </option>
43
+ </inspection_tool>
44
+ <inspection_tool class="PyShadowingBuiltinsInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false" />
45
+ <inspection_tool class="PyShadowingNamesInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false" />
46
+ <inspection_tool class="PyTypeCheckerInspection" enabled="false" level="WARNING" enabled_by_default="false" />
47
+ <inspection_tool class="PyUnresolvedReferencesInspection" enabled="true" level="WARNING" enabled_by_default="true">
48
+ <option name="ignoredIdentifiers">
49
+ <list>
50
+ <option value="dict.*" />
51
+ </list>
52
+ </option>
53
+ </inspection_tool>
54
+ <inspection_tool class="PyUnusedLocalInspection" enabled="false" level="WEAK WARNING" enabled_by_default="false" />
55
+ <inspection_tool class="ReassignedToPlainText" enabled="false" level="WARNING" enabled_by_default="false" />
56
+ <inspection_tool class="SpellCheckingInspection" enabled="false" level="TYPO" enabled_by_default="false">
57
+ <option name="processCode" value="true" />
58
+ <option name="processLiterals" value="true" />
59
+ <option name="processComments" value="true" />
60
+ </inspection_tool>
61
+ </profile>
62
+ </component>
.idea/inspectionProfiles/profiles_settings.xml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ <component name="InspectionProjectProfileManager">
2
+ <settings>
3
+ <option name="USE_PROJECT_PROFILE" value="false" />
4
+ <version value="1.0" />
5
+ </settings>
6
+ </component>
.idea/modules.xml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ <?xml version="1.0" encoding="UTF-8"?>
2
+ <project version="4">
3
+ <component name="ProjectModuleManager">
4
+ <modules>
5
+ <module fileurl="file://$PROJECT_DIR$/.idea/Game-Engine-3-LLM.iml" filepath="$PROJECT_DIR$/.idea/Game-Engine-3-LLM.iml" />
6
+ </modules>
7
+ </component>
8
+ </project>
.idea/vcs.xml ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ <?xml version="1.0" encoding="UTF-8"?>
2
+ <project version="4">
3
+ <component name="VcsDirectoryMappings">
4
+ <mapping directory="" vcs="Git" />
5
+ </component>
6
+ </project>
main.py CHANGED
@@ -4,7 +4,6 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
4
  import torch
5
  import numpy as np
6
 
7
- # Проверка версии NumPy
8
  assert np.__version__.startswith('1.'), f"Несовместимая версия NumPy: {np.__version__}"
9
 
10
  app = FastAPI()
@@ -18,21 +17,20 @@ class RequestData(BaseModel):
18
  MODEL_NAME = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
19
 
20
  try:
21
- # Загрузка модели с явным указанием device_map
22
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
23
  model = AutoModelForCausalLM.from_pretrained(
24
  MODEL_NAME,
25
  torch_dtype=torch.float32,
26
  device_map="auto",
27
- low_cpu_mem_usage=True
28
  )
29
 
30
- # Создаем pipeline без указания device
31
  generator = pipeline(
32
  "text-generation",
33
  model=model,
34
  tokenizer=tokenizer
35
  )
 
36
  except Exception as e:
37
  print(f"Ошибка загрузки модели: {str(e)}")
38
  generator = None
 
4
  import torch
5
  import numpy as np
6
 
 
7
  assert np.__version__.startswith('1.'), f"Несовместимая версия NumPy: {np.__version__}"
8
 
9
  app = FastAPI()
 
17
  MODEL_NAME = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
18
 
19
  try:
 
20
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
21
  model = AutoModelForCausalLM.from_pretrained(
22
  MODEL_NAME,
23
  torch_dtype=torch.float32,
24
  device_map="auto",
25
+ low_cpu_mem_usage=False
26
  )
27
 
 
28
  generator = pipeline(
29
  "text-generation",
30
  model=model,
31
  tokenizer=tokenizer
32
  )
33
+
34
  except Exception as e:
35
  print(f"Ошибка загрузки модели: {str(e)}")
36
  generator = None