| --- |
| license: apache-2.0 |
| language: |
| - tr |
| - en |
| base_model: SykoSLM/SykoLLM-V5.9-Mini |
| pipeline_tag: text-generation |
| library_name: transformers |
| tags: |
| - nlp |
| - code |
| - phi3 |
| - depth-up-scaling |
| - untrained |
| --- |
| |
| # SykoLLM-V6.0-Test |
|
|
| ## Model Overview |
| **SykoLLM-V6.0-Test** is an up-scaled and structurally expanded version of the previous SykoLLM models. Developed by **SykoSLM**, this model is currently in the experimental/testing phase. |
|
|
| The primary objective of this release is to provide a structurally larger foundation model by expanding both the depth (number of layers) and the width (intermediate size / MLP capacity) of the previous architecture, without losing the pre-trained knowledge. |
|
|
| ## Architectural Expansion (Up-Scaling) |
| In order to overcome the "Knowledge Interference" (capacity bottleneck) observed in previous iterations, significant architectural changes have been applied to this model: |
|
|
| * **Depth Up-Scaling (DUS):** The number of hidden layers has been increased to **24**. This was achieved by carefully duplicating and mapping the existing layers to preserve the logical and syntactic capabilities of the model. |
| * **Width Expansion (MLP Scaling):** The `intermediate_size` has been expanded to **3072**. To prevent catastrophic forgetting, the newly added weights in the feed-forward networks were initialized with exact zero (`0.0`). This ensures that the newly added parameters act as identity functions during the initial forward pass. |
|
|
| ## ⚠️ Important Notice: Status of the Model |
| **This model is currently UNTRAINED on the newly added parameters.** |
|
|
| It has been expanded solely to save pre-training time and preserve existing knowledge. While the model retains the capabilities of its predecessor, the newly added parameters (~100M+ new parameters) are currently dormant (zeroed out). |
|
|
| To fully utilize the expanded capacity and activate the new parameters, **fine-tuning is required**. If used in its current state, the model will function similarly to the previous smaller version, as the new structural capacity has not yet been fine-tuned on new or existing datasets. |
|
|
| ## Why This Approach? |
| Training a Large Language Model from scratch requires immense computational resources and time. By utilizing Net2Net (Knowledge Distillation) principles: |
| 1. We preserve the billions of tokens worth of knowledge already embedded in the model. |
| 2. We provide the model with a much larger "encyclopedic" memory (MLP expansion) to prevent data overlapping and hallucination. |
| 3. We drastically reduce the time required to achieve a higher parameter count. |
|
|
| ## Usage |
| You can load the model using the `transformers` library, but please keep in mind that it requires further fine-tuning for optimal performance. |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| model_id = "SykoSLM/SykoLLM-V6.0-Test" |
| |
| tokenizer = AutoTokenizer.from_pretrained(model_id) |
| model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto") |
| ``` |
|
|
| --- |
|
|
| Developed by SykoSLM |