| | --- |
| | language: |
| | - ur |
| | metrics: |
| | - accuracy |
| | library_name: transformers |
| | tags: |
| | - text-generation-inference |
| | --- |
| | |
| | # Model Card for Model ID |
| |
|
| | <!-- Provide a quick summary of what the model is/does. --> |
| |
|
| | This model card lists fine-tuned byT5 model for the task of Semantic Parsing. |
| |
|
| | ## Model Details |
| | We worked on a pre-trained byt5-base model and fine-tuned it with the Parallel Meaning Bank dataset (DRS-Text pairs dataset). |
| | Furthermore, we enriched the gold_silver flavors of PMB (release 5.0.0) with different augmentation strategies. |
| | |
| | |
| | ## Uses |
| | |
| | <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> |
| | |
| | To use the model, follow the code below for a quick response. |
| | |
| | ```python |
| | |
| | from transformers import ByT5Tokenizer, T5ForConditionalGeneration |
| | |
| | # Initialize the tokenizer and model |
| | tokenizer = ByT5Tokenizer.from_pretrained('saadamin2k13/urdu_semantic_parsing', max_length=512) |
| | |
| | model = T5ForConditionalGeneration.from_pretrained('saadamin2k13/urdu_semantic_parsing') |
| |
|
| | # Example sentence |
| | example = "یہ کار کالی ہے۔" |
| |
|
| | # Tokenize and prepare the input |
| | x = tokenizer(example, return_tensors='pt', padding=True, truncation=True, max_length=512)['input_ids'] |
| | |
| | # Generate output |
| | output = model.generate(x) |
| | |
| | # Decode and print the output text |
| | pred_text = tokenizer.decode(output[0], skip_special_tokens=True, clean_up_tokenization_spaces=False) |
| | print(pred_text) |
| |
|
| |
|