Meta-Llama-3.2-1B-Instruct / QCS9075 /Meta-Llama-3.2-1B-Instruct-htp.json
qc903113684's picture
Upload 7 files
73f1cb3 verified
{
"dialog" : {
"version" : 1,
"type" : "basic",
"context" : {
"version" : 1,
"size": 4096,
"n-vocab": 128256,
"bos-token": -1,
"eos-token": [128001, 128008, 128009]
},
"sampler" : {
"version" : 1,
"seed" : 42,
"temp" : 0.6,
"top-k" : 40,
"top-p" : 0.9,
"greedy" : false
},
"tokenizer" : {
"version" : 1,
"path" : "Meta-Llama-3.2-1B-Instruct-tokenizer.json"
},
"engine" : {
"version" : 1,
"n-threads" : 1,
"backend" : {
"version" : 1,
"type" : "QnnHtp",
"QnnHtp" : {
"version" : 1,
"use-mmap" : true,
"spill-fill-bufsize" : 0,
"mmap-budget" : 0,
"poll" : true,
"cpu-mask" : "0xe0",
"kv-dim" : 64,
"allow-async-init": false
},
"extensions" : "htp_backend_ext_config.json"
},
"model" : {
"version" : 1,
"type" : "binary",
"binary" : {
"version" : 1,
"ctx-bins" : [
"Meta-Llama-3.2-1B-Instruct_qnn240_qcs9075_cl4096_1_of_2.serialized.bin",
"Meta-Llama-3.2-1B-Instruct_qnn240_qcs9075_cl4096_2_of_2.serialized.bin"
]
},
"positional-encoding": {
"type": "rope",
"rope-dim": 32,
"rope-theta": 500000,
"rope-scaling": {
"rope-type": "llama3",
"factor": 32.0,
"low-freq-factor": 1.0,
"high-freq-factor": 4.0,
"original-max-position-embeddings": 8192
}
}
}
}
}
}