esteele commited on
Commit
1f4dab4
·
1 Parent(s): cbfabf3

Using try_to_load_from_cache() with repo to store model

Browse files
Files changed (1) hide show
  1. app/services/ai_service.py +3 -3
app/services/ai_service.py CHANGED
@@ -11,15 +11,15 @@ GPT2_MODEL_NAME = "distilgpt2"
11
  # os.makedirs(CACHE_DIR, exist_ok=True)
12
 
13
 
14
- for lock_file in glob.glob(os.path.join(try_to_load_from_cache(), "*.lock")):
15
  try:
16
  os.remove(lock_file)
17
  print(f"Removed lock file: {lock_file}")
18
  except Exception as e:
19
  print(f"Could not remove {lock_file}: {e}")
20
 
21
- gpt2_tokenizer = AutoTokenizer.from_pretrained(GPT2_MODEL_NAME, cache_dir=try_to_load_from_cache())
22
- gpt2_model = AutoModelForCausalLM.from_pretrained(GPT2_MODEL_NAME, cache_dir=try_to_load_from_cache())
23
 
24
  # tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
25
  # model = AutoModelWithLMHead.from_pretrained(MODEL_NAME)
 
11
  # os.makedirs(CACHE_DIR, exist_ok=True)
12
 
13
 
14
+ for lock_file in glob.glob(os.path.join(try_to_load_from_cache('funny-meme-generator', 'cache_dir'), "*.lock")):
15
  try:
16
  os.remove(lock_file)
17
  print(f"Removed lock file: {lock_file}")
18
  except Exception as e:
19
  print(f"Could not remove {lock_file}: {e}")
20
 
21
+ gpt2_tokenizer = AutoTokenizer.from_pretrained(GPT2_MODEL_NAME, cache_dir=try_to_load_from_cache('funny-meme-generator', 'cache_dir'))
22
+ gpt2_model = AutoModelForCausalLM.from_pretrained(GPT2_MODEL_NAME, cache_dir=try_to_load_from_cache('funny-meme-generator', 'cache_dir'))
23
 
24
  # tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
25
  # model = AutoModelWithLMHead.from_pretrained(MODEL_NAME)