AjaykumarPilla commited on
Commit
6d4ce31
·
verified ·
1 Parent(s): d81ff75

Update model.py

Browse files
Files changed (1) hide show
  1. model.py +13 -7
model.py CHANGED
@@ -3,7 +3,10 @@ import torch
3
  import logging
4
  from typing import Dict, List
5
  import time
6
- import psutil
 
 
 
7
 
8
  # Configure logging
9
  logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
@@ -32,9 +35,12 @@ def call_ai_model_for_insights(input_data: Dict, delay_risk: float) -> List[str]
32
  max_retries = 3
33
  retry_delay = 15 # seconds
34
 
35
- # Log system resources
36
- memory = psutil.virtual_memory()
37
- logger.info(f"System memory - Total: {memory.total / 1e9:.2f} GB, Available: {memory.available / 1e9:.2f} GB, Used: {memory.percent}%")
 
 
 
38
 
39
  for attempt in range(max_retries):
40
  try:
@@ -67,11 +73,11 @@ def call_ai_model_for_insights(input_data: Dict, delay_risk: float) -> List[str]
67
  """
68
 
69
  with torch.no_grad():
70
- inputs = tokenizer(prompt, return_tensors="pt", max_length=128, truncation=True).to("cpu")
71
  outputs = model.generate(
72
  **inputs,
73
- max_new_tokens=30,
74
- num_beams=4,
75
  temperature=0.7,
76
  do_sample=True
77
  )
 
3
  import logging
4
  from typing import Dict, List
5
  import time
6
+ try:
7
+ import psutil
8
+ except ImportError:
9
+ psutil = None
10
 
11
  # Configure logging
12
  logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
 
35
  max_retries = 3
36
  retry_delay = 15 # seconds
37
 
38
+ # Log system resources if psutil is available
39
+ if psutil:
40
+ memory = psutil.virtual_memory()
41
+ logger.info(f"System memory - Total: {memory.total / 1e9:.2f} GB, Available: {memory.available / 1e9:.2f} GB, Used: {memory.percent}%")
42
+ else:
43
+ logger.warning("psutil not available; cannot log system memory usage")
44
 
45
  for attempt in range(max_retries):
46
  try:
 
73
  """
74
 
75
  with torch.no_grad():
76
+ inputs = tokenizer(prompt, return_tensors="pt", max_length=64, truncation=True).to("cpu")
77
  outputs = model.generate(
78
  **inputs,
79
+ max_new_tokens=20,
80
+ num_beams=2,
81
  temperature=0.7,
82
  do_sample=True
83
  )