GenomeOcean-4B-GGUF

Model Overview

This is a GGUF quantized version of GenomeOcean-4B, designed for high-efficiency DNA sequence modeling.

  • Architecture: Mistral-based Genomic LLM
  • Quantization: GGUF (4-bit)
  • Primary Use: DNA sequence scoring, generation, and genomic feature analysis.

Benchmark Results (Local Evaluation)

Evaluation conducted on genomic sequences (max length 512) using TP=2.

Metric FP16 (Original) GGUF (4-bit) Change
VRAM Footprint ~2x Model Size 12.5 GB Optimized
Model Size 8112.6 MB 2443.1 MB -69.9%
NLL Loss 5.7831 5.8088 +2.61% Drift
Perplexity (PPL) 324.7482 333.2174 +2.61% Drift
Generation Time 59.5s 113.0s Optimized

Usage

Using vLLM

from vllm import LLM, SamplingParams

# Load the model
llm = LLM(model="ThomasYn/GenomeOcean-4B-GGUF")

# Generate sequences
prompts = ["ATG", "GCA"]
sampling_params = SamplingParams(temperature=0.7, top_p=0.95, max_tokens=100)
outputs = llm.generate(prompts, sampling_params)

for output in outputs:
    print(f"Generated: {output.outputs[0].text}")

Using go-infer (GenomeOcean CLI)

# Scoring sequences
python -m genomeocean.cli score --model_dir ThomasYn/GenomeOcean-4B-GGUF --sequence_file data.txt

# Generating sequences
python -m genomeocean.cli generate --model_dir ThomasYn/GenomeOcean-4B-GGUF --num 10 --max_seq_len 512

Model Repository Structure

This repository contains the necessary configuration files and model weights for GGUF inference.

  • model.gguf: Quantized weights
  • config.json: Model configuration
  • modeling_mistral.py: Architecture implementation
  • tokenizer.json & tokenizer_config.json: Genomic tokenizer files

Citation

If you use this model in your research, please cite:

@article{genomeocean2026,
  title={GenomeOcean: A Large-scale Foundation Model for Ocean Genomics},
  author={Thomas Yn, et al.},
  journal={bioRxiv},
  year={2026}
}
Downloads last month
73
GGUF
Model size
4B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support