GenomeOcean-500M-AWQ

Model Overview

This is a AWQ quantized version of GenomeOcean-500M, designed for high-efficiency DNA sequence modeling.

  • Architecture: Mistral-based Genomic LLM
  • Quantization: AWQ (4-bit)
  • Primary Use: DNA sequence scoring, generation, and genomic feature analysis.

Benchmark Results (Local Evaluation)

Evaluation conducted on genomic sequences (max length 512) using TP=2.

Metric FP16 (Original) AWQ (4-bit) Change
VRAM Footprint ~2x Model Size 4.5 GB Optimized
Model Size 1032.3 MB 286.3 MB -72.3%
NLL Loss 5.9931 6.0442 +5.24% Drift
Perplexity (PPL) 400.6407 421.6455 +5.24% Drift
Generation Time 104.1s 55.9s Optimized

Usage

Using vLLM

from vllm import LLM, SamplingParams

# Load the model
llm = LLM(model="ThomasYn/GenomeOcean-500M-AWQ", quantization="awq")

# Generate sequences
prompts = ["ATG", "GCA"]
sampling_params = SamplingParams(temperature=0.7, top_p=0.95, max_tokens=100)
outputs = llm.generate(prompts, sampling_params)

for output in outputs:
    print(f"Generated: {output.outputs[0].text}")

Using go-infer (GenomeOcean CLI)

# Scoring sequences
python -m genomeocean.cli score --model_dir ThomasYn/GenomeOcean-500M-AWQ --sequence_file data.txt

# Generating sequences
python -m genomeocean.cli generate --model_dir ThomasYn/GenomeOcean-500M-AWQ --num 10 --max_seq_len 512

Model Repository Structure

This repository contains the necessary configuration files and model weights for AWQ inference.

  • model.safetensors: Quantized weights
  • config.json: Model configuration
  • modeling_mistral.py: Architecture implementation
  • tokenizer.json & tokenizer_config.json: Genomic tokenizer files

Citation

If you use this model in your research, please cite:

@article{genomeocean2026,
  title={GenomeOcean: A Large-scale Foundation Model for Ocean Genomics},
  author={Thomas Yn, et al.},
  journal={bioRxiv},
  year={2026}
}
Downloads last month
34
Safetensors
Model size
0.5B params
Tensor type
I32
·
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support