dindizz commited on
Commit
f015c70
·
verified ·
1 Parent(s): 1a398bc

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -12
app.py CHANGED
@@ -27,35 +27,30 @@ def generate_roast(resume_text):
27
  # Define the prompt separately
28
  prompt_text = "Roast this resume:\n\n"
29
 
30
- # Tokenize the prompt to calculate its token length
31
  prompt_tokenized = tokenizer(prompt_text, return_tensors="pt")
32
  prompt_tokens = prompt_tokenized['input_ids'].shape[1]
33
 
34
- # Calculate the remaining tokens for the resume text (2048 - prompt tokens)
35
  max_resume_tokens = 2048 - prompt_tokens
36
 
37
- # Tokenize and truncate the resume text to fit within the remaining token limit
38
- resume_tokenized = tokenizer(resume_text, truncation=True, max_length=max_resume_tokens)
39
 
40
  # Decode the truncated resume back into a string
41
- truncated_resume_text = tokenizer.decode(resume_tokenized['input_ids'], skip_special_tokens=True)
42
 
43
  # Combine the prompt and the truncated resume text
44
  final_prompt = f"{prompt_text}{truncated_resume_text}\n\nRoast:"
45
 
46
- # Generate roast with the truncated prompt
47
  generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
48
 
49
- # Generate roast within max_new_tokens limit
50
  roast = generator(final_prompt, max_new_tokens=50, num_return_sequences=1)
51
 
52
  return roast[0]['generated_text']
53
 
54
-
55
-
56
-
57
-
58
-
59
  # Gradio interface function
60
  def roast_resume(file):
61
  if file.name.endswith('.pdf'):
 
27
  # Define the prompt separately
28
  prompt_text = "Roast this resume:\n\n"
29
 
30
+ # Tokenize the prompt
31
  prompt_tokenized = tokenizer(prompt_text, return_tensors="pt")
32
  prompt_tokens = prompt_tokenized['input_ids'].shape[1]
33
 
34
+ # Calculate remaining tokens for resume text (2048 - prompt tokens)
35
  max_resume_tokens = 2048 - prompt_tokens
36
 
37
+ # Tokenize and strictly truncate the resume text to fit within the remaining token space
38
+ resume_tokenized = tokenizer(resume_text, truncation=True, max_length=max_resume_tokens, return_tensors="pt")
39
 
40
  # Decode the truncated resume back into a string
41
+ truncated_resume_text = tokenizer.decode(resume_tokenized['input_ids'][0], skip_special_tokens=True)
42
 
43
  # Combine the prompt and the truncated resume text
44
  final_prompt = f"{prompt_text}{truncated_resume_text}\n\nRoast:"
45
 
46
+ # Generate the roast
47
  generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
48
 
49
+ # Ensure generated roast doesn't exceed token limit
50
  roast = generator(final_prompt, max_new_tokens=50, num_return_sequences=1)
51
 
52
  return roast[0]['generated_text']
53
 
 
 
 
 
 
54
  # Gradio interface function
55
  def roast_resume(file):
56
  if file.name.endswith('.pdf'):