Model Card for Model ID

A merged LLaMA 3.1 8B checkpoint specialized for numeric price prediction from product text. This model was created by merging LLaMA 3.1 8B with a LoRA adapter (Pricer LoRA v1) and is intended to serve as a base model for further LoRA fine-tuning.

Model Details

Overall Training Process

Training process

Model Description

pricer-merged-model-A-v1 is a transformer-based causal language model optimized for estimating approximate consumer product prices from textual metadata such as title, description, and category. This model represents a merged checkpoint (base model + LoRA), not an only-LoRA adapter.

  • Developed by: MyungHwan Hong (MightyOctopus)
  • Funded by: Self-funded / independent research
  • Shared by: MyungHwan Hong
  • Model type: Causal Language Model (Text-to-Number / Numeric Prediction)
  • Language(s) (NLP): English
  • License: MIT
  • Finetuned from model: meta-llama/Llama-3.1-8B

Model Sources [optional]

Uses

Direct Use

  • Base checkpoint for price-prediction inference

  • Base model for further LoRA fine-tuning

  • Research on LLM-based numeric regression

Downstream Use [optional]

The model is primarily intended to serve as a base checkpoint for further LoRA fine-tuning.

Possible downstream applications include:

  • Domain-specific product price predictors (e.g., electronics, books, apparel)
  • Research on LLM-based numeric regression
  • Comparative studies against classical ML regression models
  • Experiments on LoRA merging and adapter stacking strategies

Out-of-Scope Use

  • Financial decision-making

  • Legal, medical, or safety-critical applications

  • Use as an authoritative price source

Bias, Risks, and Limitations

  • Predictions are approximate, not exact

  • Performance depends on similarity to training data distribution

  • May hallucinate prices for unfamiliar products

  • Reflects historical and dataset-specific price biases

  • Not robust to rapid market price changes

Recommendations

  • Treat outputs as estimates, not ground truth

  • Validate predictions with real pricing data

  • Avoid high-stakes or commercial deployment

  • Be aware of temporal and dataset bias

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "MightyOctopus/pricer-merged-model-A-v1"
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B")

model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto"
)

prompt = """Product:
Title: Stainless Steel Electric Kettle 1.7L
Category: Home & Kitchen
Description: Fast boiling electric kettle with auto shut-off.

Price is $"""

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

with torch.no_grad():
    outputs = model.generate(inputs.input_ids, max_new_tokens=10)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

Training Data

  • Amazon product metadata

  • Fields: title, description, category, ground-truth price

  • Prices represented as structured text outputs

Training Procedure

This model was created by merging a LoRA adapter (Pricer LoRA v1) into LLaMA 3.1 8B. No additional training was performed after merging.

Training Hyperparameters

  • Training regime: bfloat16 mixed precision (inherited from LoRA training)

Evaluation

This merged checkpoint was not evaluated independently. Evaluation was only performed on downstream fine-tuned adapters (e.g. pricer-lora-ft-v3: https://huggingface.co/MightyOctopus/pricer-lora-ft-v3/blob/main/README.md).

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: NVIDIA GPU (e.g. T4, A100)
  • Hours used: ~20 hours (LoRA v1 training)
  • Cloud Provider: Google Colab / Hugging Face
  • Compute Region: Unknown
  • Carbon Emitted: Not estimated

Model Architecture and Objective

Transformer-based causal language model

Objective: Next-token prediction optimized for numeric output consistency

Hardware

NVIDIA GPU (T4, L4, A100-class)

Software

  • Transformers

  • PEFT

  • PyTorch

BibTeX:

@misc{hong2025pricermerged, author = {MyungHwan Hong}, title = {Pricer Merged LLaMA 3.1 8B Model}, year = {2025}, url = {https://huggingface.co/MightyOctopus/pricer-merged-model-A-v1} }

APA:

MyungHwan Hong. (2025). Pricer Merged LLaMA 3.1 8B Model. Hugging Face.

Model Card Authors [optional]

MyungHwan Hong

Model Card Contact

Hugging Face: MightyOctopus

Downloads last month
31
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MightyOctopus/pricer-merged-model-A-v1

Finetuned
(1768)
this model
Adapters
1 model

Space using MightyOctopus/pricer-merged-model-A-v1 1

Paper for MightyOctopus/pricer-merged-model-A-v1