dindizz commited on
Commit
e0548b6
·
verified ·
1 Parent(s): b1dd7a1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -9
app.py CHANGED
@@ -27,19 +27,23 @@ def generate_roast(resume_text):
27
  # Define the prompt separately
28
  prompt_text = "Roast this resume:\n\n"
29
 
30
- # Tokenize the prompt and resume text, truncating to fit within the 2048 token limit
31
- max_tokens = 2048 - len(tokenizer(prompt_text)['input_ids']) # Reserve space for the prompt
32
- inputs = tokenizer(resume_text, return_tensors="pt", truncation=True, max_length=max_tokens)
33
-
34
- # Convert the tokenized inputs back to text for the prompt
35
- truncated_resume_text = tokenizer.decode(inputs["input_ids"][0], skip_special_tokens=True)
 
 
 
 
 
36
 
37
- # Create the final prompt by combining the prompt and the truncated resume
38
  prompt = f"{prompt_text}{truncated_resume_text}\n\nRoast:"
39
-
40
- generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
41
 
42
  # Generate roast
 
43
  roast = generator(prompt, max_new_tokens=50, num_return_sequences=1)
44
 
45
  return roast[0]['generated_text']
@@ -47,6 +51,7 @@ def generate_roast(resume_text):
47
 
48
 
49
 
 
50
  # Gradio interface function
51
  def roast_resume(file):
52
  if file.name.endswith('.pdf'):
 
27
  # Define the prompt separately
28
  prompt_text = "Roast this resume:\n\n"
29
 
30
+ # Calculate how many tokens the prompt uses
31
+ prompt_tokens = tokenizer(prompt_text, return_tensors="pt")['input_ids'].shape[1]
32
+
33
+ # Ensure the total length (prompt + resume text) doesn't exceed 2048 tokens
34
+ max_resume_tokens = 2048 - prompt_tokens
35
+
36
+ # Tokenize the resume text and truncate to max_resume_tokens
37
+ resume_tokens = tokenizer(resume_text, return_tensors="pt", truncation=True, max_length=max_resume_tokens)
38
+
39
+ # Decode the truncated resume back into text
40
+ truncated_resume_text = tokenizer.decode(resume_tokens['input_ids'][0], skip_special_tokens=True)
41
 
42
+ # Create the final prompt with the truncated resume text
43
  prompt = f"{prompt_text}{truncated_resume_text}\n\nRoast:"
 
 
44
 
45
  # Generate roast
46
+ generator = pipeline('text-generation', model=model, tokenizer=tokenizer)
47
  roast = generator(prompt, max_new_tokens=50, num_return_sequences=1)
48
 
49
  return roast[0]['generated_text']
 
51
 
52
 
53
 
54
+
55
  # Gradio interface function
56
  def roast_resume(file):
57
  if file.name.endswith('.pdf'):