Mfusenig commited on
Commit
96bf2b2
·
verified ·
1 Parent(s): 69a718a

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +33 -33
  2. handler.py +3 -2
  3. requirements.txt +4 -5
README.md CHANGED
@@ -1,33 +1,33 @@
1
- ---
2
- pipeline_tag: text-generation
3
- library_name: transformers
4
- license: apache-2.0
5
- base_model: google/t5gemma-s-s-ul2-it
6
- model_type: t5gemma
7
- ---
8
-
9
- # T5Gemma Fine-tuned Model
10
-
11
- This is a fine-tuned T5Gemma model for text-to-text generation tasks.
12
-
13
- ## Model Details
14
-
15
- - **Base Model**: google/t5gemma-s-s-ul2-it
16
- - **Architecture**: T5GemmaForConditionalGeneration
17
- - **Task**: Text-to-text generation
18
- - **Framework**: Transformers
19
-
20
- ## Usage
21
-
22
- ```python
23
- from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
24
-
25
- tokenizer = AutoTokenizer.from_pretrained("your-username/model-name")
26
- model = AutoModelForSeq2SeqLM.from_pretrained("your-username/model-name")
27
-
28
- # Use with chat template
29
- messages = [{"role": "user", "content": "Your input text here"}]
30
- input_ids = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
31
- outputs = model.generate(input_ids, max_new_tokens=1024, temperature=0.1, do_sample=True)
32
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
33
- ```
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ library_name: transformers
4
+ license: apache-2.0
5
+ base_model: google/t5gemma-s-s-ul2-it
6
+ model_type: t5gemma
7
+ ---
8
+
9
+ # T5Gemma Fine-tuned Model
10
+
11
+ This is a fine-tuned T5Gemma model for text-to-text generation tasks.
12
+
13
+ ## Model Details
14
+
15
+ - **Base Model**: google/t5gemma-s-s-ul2-it
16
+ - **Architecture**: T5GemmaForConditionalGeneration
17
+ - **Task**: Text-to-text generation
18
+ - **Framework**: Transformers
19
+
20
+ ## Usage
21
+
22
+ ```python
23
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
24
+
25
+ tokenizer = AutoTokenizer.from_pretrained("your-username/model-name")
26
+ model = AutoModelForSeq2SeqLM.from_pretrained("your-username/model-name")
27
+
28
+ # Use with chat template
29
+ messages = [{"role": "user", "content": "Your input text here"}]
30
+ input_ids = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt")
31
+ outputs = model.generate(input_ids, max_new_tokens=1024, temperature=0.1, do_sample=True)
32
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True)
33
+ ```
handler.py CHANGED
@@ -6,7 +6,8 @@ class EndpointHandler:
6
  self.tokenizer = AutoTokenizer.from_pretrained(path)
7
  self.model = AutoModelForSeq2SeqLM.from_pretrained(
8
  path,
9
- torch_dtype=torch.bfloat16
 
10
  )
11
 
12
  def __call__(self, data):
@@ -28,4 +29,4 @@ class EndpointHandler:
28
 
29
  return {
30
  "generated_text": self.tokenizer.decode(outputs[0], skip_special_tokens=True)
31
- }
 
6
  self.tokenizer = AutoTokenizer.from_pretrained(path)
7
  self.model = AutoModelForSeq2SeqLM.from_pretrained(
8
  path,
9
+ torch_dtype=torch.bfloat16,
10
+ trust_remote_code=True # Allow loading of custom model architectures
11
  )
12
 
13
  def __call__(self, data):
 
29
 
30
  return {
31
  "generated_text": self.tokenizer.decode(outputs[0], skip_special_tokens=True)
32
+ }
requirements.txt CHANGED
@@ -1,5 +1,4 @@
1
- torch==2.6
2
- torchvision==0.16.2
3
- transformers==4.54.1
4
- sentencepiece>=0.1.99
5
- accelerate>=0.21.0
 
1
+ torch>=2.4.0
2
+ transformers==4.54.1
3
+ sentencepiece>=0.1.99
4
+ accelerate>=0.21.0