Text2MCDM
Collection
Natural language to Z-number MCDM pipeline • 4 items • Updated
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("nuriyev/text2mcdm")
model = AutoModelForCausalLM.from_pretrained("nuriyev/text2mcdm")
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))This model extracts structured Z-number decision matrices from conversational text describing multi-criteria decision problems. Given a natural language narrative about alternatives, criteria, and preferences (often messy, subjective, or contradictory), the model outputs a markdown table with:
value:confidence format (e.g., 4:3 = good rating with moderate confidence)Z-numbers extend traditional fuzzy numbers by incorporating reliability/confidence, making them ideal for real-world decision-making under uncertainty.
The extracted matrix can be analyzed using Z-number-based MCDM methods (TOPSIS, PROMETHEE) to produce ranked alternatives. See text2mcdm for the full pipeline.
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="nuriyev/text2mcdm") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)