| | --- |
| | license: mit |
| | language: |
| | - en |
| | base_model: |
| | - distilbert/distilgpt2 |
| | library_name: transformers |
| | tags: |
| | - text-generation-inference |
| | - words |
| | - text2gpt |
| | --- |
| | # Text2GPT (81.9M parameters) |
| | Currently Text2GPT uses the base model: distilbert/distilgpt2 to fine-tune |
| |
|
| | # Files |
| | The following JSON files here: |
| | - tokenizer_config.json |
| | ```json |
| | { |
| | "add_bos_token": false, |
| | "add_prefix_space": false, |
| | "added_tokens_decoder": { |
| | "50256": { |
| | "content": "<|endoftext|>", |
| | "lstrip": false, |
| | "normalized": true, |
| | "rstrip": false, |
| | "single_word": false, |
| | "special": true |
| | } |
| | }, |
| | "bos_token": "<|endoftext|>", |
| | "clean_up_tokenization_spaces": false, |
| | "eos_token": "<|endoftext|>", |
| | "errors": "replace", |
| | "extra_special_tokens": {}, |
| | "model_max_length": 1024, |
| | "pad_token": "<|endoftext|>", |
| | "tokenizer_class": "GPT2Tokenizer", |
| | "unk_token": "<|endoftext|>" |
| | } |
| | ``` |
| | - config.json |
| | ```json |
| | { |
| | "_num_labels": 1, |
| | "activation_function": "gelu_new", |
| | "architectures": [ |
| | "GPT2LMHeadModel" |
| | ], |
| | "attn_pdrop": 0.1, |
| | "bos_token_id": 50256, |
| | "embd_pdrop": 0.1, |
| | "eos_token_id": 50256, |
| | "id2label": { |
| | "0": "LABEL_0" |
| | }, |
| | "initializer_range": 0.02, |
| | "label2id": { |
| | "LABEL_0": 0 |
| | }, |
| | "layer_norm_epsilon": 1e-05, |
| | "model_type": "gpt2", |
| | "n_ctx": 1024, |
| | "n_embd": 768, |
| | "n_head": 12, |
| | "n_inner": null, |
| | "n_layer": 6, |
| | "n_positions": 1024, |
| | "reorder_and_upcast_attn": false, |
| | "resid_pdrop": 0.1, |
| | "scale_attn_by_inverse_layer_idx": false, |
| | "scale_attn_weights": true, |
| | "summary_activation": null, |
| | "summary_first_dropout": 0.1, |
| | "summary_proj_to_labels": true, |
| | "summary_type": "cls_index", |
| | "summary_use_proj": true, |
| | "task_specific_params": { |
| | "text-generation": { |
| | "do_sample": true, |
| | "max_length": 50 |
| | } |
| | }, |
| | "torch_dtype": "float32", |
| | "transformers_version": "4.50.3", |
| | "use_cache": true, |
| | "vocab_size": 50257 |
| | } |
| | ``` |
| | other files... |
| | # Use it: |
| | ## Load model directly |
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | |
| | tokenizer = AutoTokenizer.from_pretrained("kulia-moon/Text2GPT") |
| | model = AutoModelForCausalLM.from_pretrained("kulia-moon/Text2GPT") |
| | ``` |
| | ## Use a pipeline as a high-level helper |
| | ```python |
| | from transformers import pipeline |
| |
|
| | pipe = pipeline("text-generation", model="kulia-moon/Text2GPT") |
| | ``` |
| | # vLLM use: |
| | ## Deploy with docker on Linux: |
| | ```shell |
| | docker run --runtime nvidia --gpus all \ |
| | --name my_vllm_container \ |
| | -v ~/.cache/huggingface:/root/.cache/huggingface \ |
| | --env "HUGGING_FACE_HUB_TOKEN=<secret>" \ |
| | -p 8000:8000 \ |
| | --ipc=host \ |
| | vllm/vllm-openai:latest \ |
| | # --model kulia-moon/Text2GPT |
| | ``` |
| | ## Load and run the model: |
| | ```shell |
| | docker exec -it my_vllm_container bash -c "vllm serve kulia-moon/Text2GPT" |
| | ``` |