Instructions to use ansukla/task-llm with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use ansukla/task-llm with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="ansukla/task-llm")# Load model directly from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("ansukla/task-llm") model = AutoModelForSeq2SeqLM.from_pretrained("ansukla/task-llm") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use ansukla/task-llm with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "ansukla/task-llm" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ansukla/task-llm", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/ansukla/task-llm
- SGLang
How to use ansukla/task-llm with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "ansukla/task-llm" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ansukla/task-llm", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "ansukla/task-llm" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "ansukla/task-llm", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use ansukla/task-llm with Docker Model Runner:
docker model run hf.co/ansukla/task-llm
YAML Metadata Warning:The pipeline tag "text2text-generation" is not in the official list: text-classification, token-classification, table-question-answering, question-answering, zero-shot-classification, translation, summarization, feature-extraction, text-generation, fill-mask, sentence-similarity, text-to-speech, text-to-audio, automatic-speech-recognition, audio-to-audio, audio-classification, audio-text-to-text, voice-activity-detection, depth-estimation, image-classification, object-detection, image-segmentation, text-to-image, image-to-text, image-to-image, image-to-video, unconditional-image-generation, video-classification, reinforcement-learning, robotics, tabular-classification, tabular-regression, tabular-to-text, table-to-text, multiple-choice, text-ranking, text-retrieval, time-series-forecasting, text-to-video, image-text-to-text, image-text-to-image, image-text-to-video, visual-question-answering, document-question-answering, zero-shot-image-classification, graph-ml, mask-generation, zero-shot-object-detection, text-to-3d, image-to-3d, image-feature-extraction, video-text-to-text, keypoint-detection, visual-document-retrieval, any-to-any, video-to-video, other
Model Card for task-llm
This model supports abstractive QA tasks. Given a set of passages and a question, it tries to generate a comprehensive answer by reading the passages. In other words, the model does the generation part of retrieval augmented generation (RAG).
Model Details
This model was intended to be a T5 style multi task model trained with Bart to leverage the larger context length and better performance. At the moment, the only task supported by this model is abstractive qa
Model Description
- Developed by: Ambika Sukla, Nlmatics Corp.
- Model type: Generative Language Model, Abstractive QA, QASum
- Language(s) (NLP): English
- License: Apache 2.0
- Finetuned from model bart:
Uses
This model supports abstractive QA tasks. Given a set of passages and a question, it tries to generate a comprehensive answer by reading the passages.
Bias, Risks, and Limitations
This model is trained with a very simple dataset and will need further fine tuning for your use cases.
Recommendations
Fine tune the model with your own data.
How to Get Started with the Model
Use the following prompt: prompt = f"###Task: abstractive_qa \n###Question: {question} \n###Passages:{passage}"
where question is your query and passage is a concatenated set of passages that needs to be considered for answering a question.
Use the code below to get started with the model:
To run this code with nlm-model-service, use the following code:
pip install nlm-utils
qa_sum_client_bart = ClassificationClient(
model="bart",
task="qa_sum",
url=v100Url,
retry=1,
)
# nlm-model-service suppports batch invocation and you can send multiple question/passage pairs at a time.
questions = ["what are the adverse reactions of Dimethylsulfoxide"]
sentences = ["Dimethylsulfoxide Adverse reactions Garlic taste in mouth, dry skin, erythema and pruritis (2), urine discoloration, halitosis, agitation, hypotension, sedation and dizziness (13) have been reported following use of DMSO. Dimethylsulfoxide Adverse reactions: malaria and loose motion."]
qa_sum_client_bart(questions, sentences)
Training Details
Training Data
Base training data was taken from this dataset with more data added for certain usage scenarios. https://github.com/microsoft/MSMARCO-Question-Answering
Training Procedure
Coming soon.
Hardware
T4, V100 or A100 GPU is recommended.
Citation
MS MARCO: A Human Generated MAchine Reading COmprehension Dataset https://arxiv.org/abs/1611.09268
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension https://arxiv.org/abs/1910.13461
Model Card Authors
Ambika Sukla
Model Card Contact
- Downloads last month
- 16
docker model run hf.co/ansukla/task-llm