CrashOverrideX commited on
Commit
175d37c
·
verified ·
1 Parent(s): c0e6da2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +53 -33
README.md CHANGED
@@ -1,33 +1,53 @@
1
-
2
- # Quillan-Ronin Llama-Compatible GGUF Export
3
-
4
- ## Files Created:
5
- - model.safetensors: Model weights in Llama-compatible format (float16)
6
- - config.json: Llama model configuration
7
- - tokenizer_config.json: Tokenizer configuration
8
- - vocab.json: Vocabulary mapping
9
- - special_tokens_map.json: Special tokens mapping
10
-
11
- ## GGUF Conversion Instructions:
12
-
13
- ### Option 1: Local Conversion (Recommended)
14
- ```bash
15
- cd llama.cpp
16
- python convert_hf_to_gguf.py ../gguf_export/ --outtype f16
17
- ./build/bin/llama-quantize model.gguf model-Q4_K_M.gguf Q4_K_M
18
- ```
19
-
20
- ### Option 2: Online Conversion
21
- Upload gguf_export folder to: https://huggingface.co/spaces/ggml-org/gguf-my-repo
22
- Select GGUF output format and Q4_K_M quantization
23
-
24
- ### Option 3: Direct Ollama Creation
25
- ```bash
26
- ollama create quillan-ronin -f ./gguf_export
27
- ```
28
-
29
- ## Architecture Notes:
30
- - Uses LlamaForCausalLM architecture for compatibility
31
- - Maintains Quillan's core capabilities through weight mapping
32
- - 4 transformer layers, 8 attention heads, 1024 hidden dim
33
- - Supports standard Llama inference pipelines
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language: en
4
+ library_name: transformers
5
+ tags:
6
+ - llama
7
+ - causal-lm
8
+ - multimodal
9
+ - text-generation
10
+ - image-generation
11
+ - video-generation
12
+ - audio-generation
13
+ ---
14
+
15
+ # Quillan-Ronin v6.0 - Expanded Council Architecture
16
+
17
+ An advanced multimodal AI with 34 expert councils and 300,000 micro-agents.
18
+
19
+ ## Model Description
20
+
21
+ Quillan-Ronin v6.0 is an expanded multimodal AI system featuring:
22
+ - 4K resolution image generation
23
+ - 720p 30fps video generation
24
+ - 10-minute high-quality audio generation
25
+ - 34 expert councils for specialized reasoning
26
+ - 300,000 micro-agent swarm controller
27
+ - 12 vector processing cores
28
+
29
+ ## Architecture
30
+
31
+ - **Model Type**: LlamaForCausalLM (GGUF compatible)
32
+ - **Parameters**: ~68M (lightweight version)
33
+ - **Context Length**: 1024 tokens
34
+ - **Hidden Size**: 1024
35
+ - **Layers**: 4 transformer layers
36
+ - **Attention Heads**: 8
37
+
38
+ ## Capabilities
39
+
40
+ - Text processing and generation
41
+ - Image generation (4K resolution)
42
+ - Video generation (720p @ 30fps)
43
+ - Audio generation (44.1kHz, 10-minute clips)
44
+ - Multimodal integration
45
+ - Advanced reasoning through council system
46
+
47
+ ## Usage
48
+
49
+ This model is designed for GGUF conversion and use with Ollama or llama.cpp.
50
+
51
+ ## License
52
+
53
+ MIT License