| | --- |
| | license: apache-2.0 |
| | --- |
| | |
| |
|
| |
|
| | Nikhitha Telugu Dataset Model |
| | Model ID: Nikitha-logics/Nikhitha_telugu_dataset_model |
| | |
| | Model Type: ALBERT-based Language Model |
| | |
| | License: Apache-2.0 |
| | |
| | Model Overview |
| | The Nikhitha Telugu Dataset Model is an ALBERT-based language model trained on a Telugu language dataset. ALBERT (A Lite BERT) is a transformer-based model designed for natural language processing tasks, optimized for efficiency and performance. |
| | |
| | Model Details |
| | Model Size: 33.2 million parameters |
| | |
| | Tensor Type: Float32 (F32) |
| | |
| | Format: Safetensors |
| | |
| | Usage |
| | To utilize this model in your projects, you can load it using the Hugging Face Transformers library: |
| | |
| | from transformers import AlbertTokenizer, AlbertForMaskedLM |
| | |
| | # Load the tokenizer |
| | tokenizer = AlbertTokenizer.from_pretrained("Nikitha-logics/Nikhitha_telugu_dataset_model") |
| | |
| | # Load the model |
| | model = AlbertForMaskedLM.from_pretrained("Nikitha-logics/Nikhitha_telugu_dataset_model") |
| | |