Instructions to use sh-024/LFM2.5-1.2B-Instruct-Korean with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use sh-024/LFM2.5-1.2B-Instruct-Korean with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="sh-024/LFM2.5-1.2B-Instruct-Korean") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("sh-024/LFM2.5-1.2B-Instruct-Korean") model = AutoModelForCausalLM.from_pretrained("sh-024/LFM2.5-1.2B-Instruct-Korean") messages = [ {"role": "user", "content": "Who are you?"}, ] inputs = tokenizer.apply_chat_template( messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt", ).to(model.device) outputs = model.generate(**inputs, max_new_tokens=40) print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:])) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use sh-024/LFM2.5-1.2B-Instruct-Korean with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "sh-024/LFM2.5-1.2B-Instruct-Korean" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "sh-024/LFM2.5-1.2B-Instruct-Korean", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker
docker model run hf.co/sh-024/LFM2.5-1.2B-Instruct-Korean
- SGLang
How to use sh-024/LFM2.5-1.2B-Instruct-Korean with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "sh-024/LFM2.5-1.2B-Instruct-Korean" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "sh-024/LFM2.5-1.2B-Instruct-Korean", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "sh-024/LFM2.5-1.2B-Instruct-Korean" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "sh-024/LFM2.5-1.2B-Instruct-Korean", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' - Unsloth Studio new
How to use sh-024/LFM2.5-1.2B-Instruct-Korean with Unsloth Studio:
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for sh-024/LFM2.5-1.2B-Instruct-Korean to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex # Run unsloth studio unsloth studio -H 0.0.0.0 -p 8888 # Then open http://localhost:8888 in your browser # Search for sh-024/LFM2.5-1.2B-Instruct-Korean to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required # Open https://huggingface.co/spaces/unsloth/studio in your browser # Search for sh-024/LFM2.5-1.2B-Instruct-Korean to start chatting
Load model with FastModel
pip install unsloth from unsloth import FastModel model, tokenizer = FastModel.from_pretrained( model_name="sh-024/LFM2.5-1.2B-Instruct-Korean", max_seq_length=2048, ) - Docker Model Runner
How to use sh-024/LFM2.5-1.2B-Instruct-Korean with Docker Model Runner:
docker model run hf.co/sh-024/LFM2.5-1.2B-Instruct-Korean
LFM2.5-1.2B-Instruct-Korean
Model Overview
LFM2.5-1.2B-Instruct-Korean is a Korean instruction-following language model based on LiquidAI/LFM2.5-1.2B-Instruct.
This model was fine-tuned on Korean-centered datasets with the goal of improving performance on Korean question answering, general conversation, and instruction-following tasks.
The model is designed to generate responses that are more natural, consistent, and contextually appropriate in Korean.
Base Model
- Base model:
LiquidAI/LFM2.5-1.2B-Instruct
Training Data
This model was fine-tuned using the following Korean-centered datasets:
maywell/koVastCarrotAI/ko-instruction-datasetMarkrAI/KoCommercial-Dataset
The training data includes Korean instruction-response pairs, general conversational data, and commercially or practically oriented text.
This setup was intended to help the model learn a broad range of Korean expressions, styles, and contexts.
Training Details
The model was trained for 1 full epoch over the entire dataset.
Training proceeded for a total of 25,652 steps and was completed successfully without interruption.
Training Summary
- Number of epochs: 1
- Total training steps: 25,652
- Training completion status: Completed
- Training status: Stable
Evaluation Results
The final evaluation metrics are as follows:
- training_loss: 1.1046
- eval_loss: 1.1015
- eval_perplexity: 3.0087
These results show that the training loss and evaluation loss remained at nearly the same level, suggesting that the model demonstrated relatively stable generalization performance on the validation set without a clear sign of overfitting.
Result Interpretation
One notable point in this experiment is the very small gap between training loss and evaluation loss.
- training_loss = 1.1046
- eval_loss = 1.1015
In general, a large gap between these two values may indicate overfitting.
However, in this experiment, the difference was very small, which suggests that the model adapted to the training data in a stable manner while maintaining a similar level of performance on the validation set.
In addition, the result of eval_perplexity = 3.0087 indicates that the model predicts the next token in a relatively stable and consistent way.
Taken together, these results suggest that the model successfully learned the major patterns in the training data and converged stably without severe instability.
Overall Conclusion
This fine-tuning was completed successfully and showed overall solid results.
The close alignment between training loss and evaluation loss suggests that the model did not significantly overfit during training and that the optimization process remained stable.
Moreover, no sharp divergence in loss values was observed throughout the training process.
Therefore, this experiment can be regarded as a fine-tuning run that converged stably overall.
Limitations and Future Work
While these metrics are useful for assessing training stability and convergence, they do not fully reflect response quality, factuality, instruction-following accuracy, or real-world usability.
To obtain a more comprehensive evaluation of the model, the following additional assessments are planned:
- evaluation on real user question-answer examples
- downstream task performance evaluation
- qualitative analysis of generated responses
- safety and hallucination checks
Through these follow-up evaluations, we aim to verify whether the model can go beyond stable training-time metrics and provide reliable and consistent performance in real-world usage scenarios.
License
This model is a fine-tuned derivative of LiquidAI/LFM2.5-1.2B-Instruct.
Use and distribution of this model are subject to the terms of the LFM Open License v1.0 applicable to the base model.
Please also review any additional obligations arising from the datasets used during fine-tuning.
Feedback
This model will continue to be improved through further evaluation and refinement.
If you have any feedback on your experience using the model or notice areas that need improvement, your input will be carefully considered and reflected in future quality improvements.
Your feedback will be a great source of support in improving and further developing the model.
- Downloads last month
- 31