How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="llmware/slim-summary-phi-3", trust_remote_code=True)
messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("llmware/slim-summary-phi-3", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("llmware/slim-summary-phi-3", trust_remote_code=True)
messages = [
    {"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
	messages,
	add_generation_prompt=True,
	tokenize=True,
	return_dict=True,
	return_tensors="pt",
).to(model.device)

outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
Quick Links

slim-summary-phi-3

slim-summary-phi-3 is a specialized function calling model that summarizes a given text and generates as output a Python list of summary points.

This is the base Pytorch version of the model, useful for further fine-tuning. For faster inference, we would recommend using either the GGUF or OpenVino version of the model.

Model Description

  • Developed by: llmware
  • Model type: phi-3
  • Parameters: 3.8 billion
  • Language(s) (NLP): English
  • License: Apache 2.0
  • Uses: Summary bulletpoints extracted from complex business documents

Model Card Contact

llmware on github

llmware on hf

llmware website

Downloads last month
11
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for llmware/slim-summary-phi-3

Quantizations
2 models

Collection including llmware/slim-summary-phi-3