How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="OpenOneRec/OneRec-1.7B-pretrain")
messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("OpenOneRec/OneRec-1.7B-pretrain")
model = AutoModelForCausalLM.from_pretrained("OpenOneRec/OneRec-1.7B-pretrain")
messages = [
    {"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
	messages,
	add_generation_prompt=True,
	tokenize=True,
	return_dict=True,
	return_tensors="pt",
).to(model.device)

outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
Quick Links

OpenOneRec

An Open Foundation Model and Benchmark to Accelerate Generative Recommendation

Hugging Face GitHub Code Paper License


πŸ“– OneRec-Foundation-Pretrain Models

This repository provides the pre-trained weights of the OneRec-Foundation series, which has undergone Itemic-Text Alignment and Full-Parameter Co-Pretraining.

We release this checkpoint to enable users to perform customized post-training or alignment tailored to their specific downstream tasks and datasets, providing greater flexibility for specialized research.

For technical details on the pre-training architecture, please refer to our Technical Report.

Downloads last month
31
Safetensors
Model size
2B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Paper for OpenOneRec/OneRec-1.7B-pretrain