midorin-Linux/plamo-3-nict-8b-base-bnb-4bit

This model is a bnb 4bit version of pfnet/plamo-3-nict-8b-base.

Base model: pfnet/plamo-3-nict-8b-base

Model Description

PLaMo 3 NICT 8B Base is a 8B model pre-trained on English and Japanese datasets, developed by Preferred Networks, Inc. collaborative with National Institute of Information and Communications Technology, NICT.

PLaMo 3 NICT models adapt a hybrid architecture with Sliding Window Attention (SWA) and Traditional Attetntion layers.

PLaMo 3 NICT 8B Base is released under PLaMo community license. Please check the following license and agree to this before downloading.

NOTE: This model has NOT been instruction-tuned for chat dialog or other downstream tasks.

For commercial users

Please check the PLaMo community license and contact us via the following form to use commercial purpose.

Usage

Requirements

python>=3.10.0
numpy>=1.26.4
numba>=0.60.0
torch>=2.6.0,<=2.9.0
transformers>=4.56.0,<=4.57.1

Use vLLM

We provide the implementation via vLLM plugin.

See more details to setup the plugin for PLaMo 3 models: https://github.com/pfnet-research/vllm-plamo3-plugin

import vllm
model = vllm.LLM("midorin-Linux/plamo-3-nict-8b-base-bnb-4bit", trust_remote_code=True)
response = model.generate(
    prompts=["The future of artificial intelligence technology is "],
    sampling_params=vllm.SamplingParams(
        n=1,
        max_tokens=32,
        top_k=50,
        top_p=0.95,
        temperature=1.0,
    ),
)[0]
generated_text = response.outputs[0].text
print(generated_text)

Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("midorin-Linux/plamo-3-nict-8b-base-bnb-4bit", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("midorin-Linux/plamo-3-nict-8b-base-bnb-4bit", trust_remote_code=True)
text = "銇撱倢銇嬨倝銇汉宸ョ煡鑳芥妧琛撱伅"
input_ids = tokenizer(text, return_tensors="pt").input_ids
generated_tokens = model.generate(
    inputs=input_ids,
    max_new_tokens=32,
    do_sample=True,
    top_k=50,
    top_p=0.95,
    temperature=1.0,
)[0]
generated_text = tokenizer.decode(generated_tokens)
print(generated_text)

Model Details

  • Model size: 8B
  • Trained tokens: 800B
  • Developed by:
    • Preferred Networks, Inc.
    • collaborative with National Institute of Information and Communications Technology, NICT
  • Model type: Causal decoder-only
  • Language(s): English, Japanese
  • License: PLaMo community license

Training Dataset

We trained PLaMo 3 NICT 8B Base in two phases, phase 1 with 875B tokens and phase 2 with 125B tokens. The percentage of datasets in each phase is shown in the following table.

875B (phase 1) 125B (phase 2) Tokens
English 45% 35% 437.5B
Japanese 30% 40% 312.5B
Code 15% 15% 150.0B
Other 10% 10% 100.0B

The training dataset also includes the following publicly available community resources:

Tokenizer

PLaMo 3 NICT 8B Base tokenizer is optimized by numba, which is JIT compiler for numerical functions. The tokenizer is trained on a subst of the datasets for model pre-training.

Bias, Risks, and Limitations

PLaMo 3 NICT 8B Base is a new technology that carries risks with use. Testing conducted to date has been in English and Japanese, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, PLaMo 3 NICT 8B Base's potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of PLaMo 3 NICT 8B Base, developers should perform safety testing and tuning tailored to their specific applications of the model.

AI policies for Preferred Networks, Inc. group

Downloads last month
137
Safetensors
Model size
8B params
Tensor type
F32
F16
U8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for midorin-Linux/plamo-3-nict-8b-base-bnb-4bit

Quantized
(2)
this model