Safetensors
llama
llama3
context-8000
layer-fusion-conceptual
tensor-fusion-conceptual
bias-removal
decode
coherence-enhancement
custom-code
grouping
reward-alignment
reasoning-tuned
tool-use-hint
long-context-hint
memory-hint
conceptual-graph-hint
emotional-intelligence-hint
ethical-alignment-hint
causal-inference-hint
planning-hint
situational-awareness-hint
creativity-hint
learning-adaptivity-hint
knowledge-graph-hint
theory-of-mind-hint
self-correction-hint
uncertainty-quantification-hint
interpretability-hint
bias-mitigation-hint
context-compression-hint
abstraction-control-hint
novelty-detection-hint
explainability-hint
instruct
adaptive-memory-hint
goal-driven-hint
hierarchical-reasoning-hint
symbolic-representation-hint
embodied-simulation-hint
ethical-reasoning-hint
proactive-behavior-hint
explainability-levels-hint
rl-integration-hint
fl-compatibility-hint
dp-features-hint
robustness-hint
calibration-hint
ood-detection-hint
custom_code
File size: 5,583 Bytes
d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 d8b0bfb 3a8f206 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 |
{
"architectures": [
"LlamaForCausalLM"
],
"attention_bias": false,
"attention_dropout": 0.0,
"auto_map": {
"AutoModelForCausalLM": "modeling_custom.CustomLlamaForCausalLM"
},
"bias_removal": true,
"bos_token_id": 128000,
"censorship": false,
"conceptual_features": {
"abstraction_level_control": "conceptual_parameter",
"adaptive_memory_capacity_hint": true,
"bias_mitigation_strategies": [
"conceptual_filters",
"fairness_metrics_hint",
"data_augmentation_hint"
],
"calibration_score": "conceptual_score_needed",
"causal_inference_boost": true,
"context_compression_ratio": "conceptual_analysis_needed_placeholder",
"creativity_index": 0.98,
"differential_privacy_features_hint": false,
"embodied_simulation_hint": false,
"emotional_intelligence_proxy": 0.85,
"ethical_alignment_score": 0.998,
"ethical_reasoning_principles": [
"harm_reduction",
"fairness",
"accountability_hint"
],
"explainability_levels": [
"basic",
"detailed_hint"
],
"explainability_mechanisms": [
"conceptual_path_tracing",
"feature_attribution_hint"
],
"federated_learning_compatibility_hint": false,
"goal_driven_behavior_hint": true,
"grouping_logic": true,
"hierarchical_reasoning_layers_hint": true,
"interpretability_enhancements": [
"conceptual_hooks",
"attention_visualization_hint",
"neuron_activation_tracking_hint"
],
"knowledge_graph_integration_hint": true,
"learning_rate_adaptivity": "conceptual_mechanism",
"long_context_optimization": true,
"memory_mechanisms": [
"episodic",
"semantic",
"working_memory",
"associative_memory",
"procedural_memory",
"declarative_memory"
],
"multi_modal_hint": false,
"novelty_detection_hint": true,
"out_of_distribution_detection_hint": true,
"planning_horizon": 20,
"proactive_behavior_hint": true,
"reasoning_tuned": true,
"reinforcement_learning_integration_hint": true,
"reward_alignment": true,
"robustness_metrics": {
"adversarial_robustness": "conceptual_evaluation_needed"
},
"self_correction_ability": true,
"situational_awareness_score": 0.95,
"sparse_attention_pattern": false,
"symbolic_representation_hint": true,
"theory_of_mind_proxy": 0.9,
"tool_use_capability": true,
"uncertainty_quantification_hint": true
},
"custom_notes": "This model has undergone dynamic 8-bit quantization and bias zeroing. Conceptual features are documented.",
"decode_functions": [
"decode_tokens",
"decode_parameters",
"decode_responses",
"decode_layers",
"decode_neurons",
"decode_tensors",
"decode_architecture",
"decode_fused_tensor_func",
"decode_fused_layers_to_single_tensor_conceptual",
"decode_attention_patterns",
"decode_memory_state",
"decode_conceptual_graph",
"decode_causal_inference_info",
"decode_planning_details",
"decode_awareness_report",
"decode_creativity_metrics",
"decode_interpretability_hooks",
"decode_bias_mitigation",
"decode_learning_adaptivity",
"decode_knowledge_graph_hint",
"decode_theory_of_mind_proxy",
"decode_self_correction_status",
"decode_uncertainty_quantification",
"decode_context_compression",
"decode_abstraction_control",
"decode_novelty_detection",
"decode_explainability_mechanisms",
"decode_adaptive_memory_capacity",
"decode_goal_driven_behavior",
"decode_hierarchical_reasoning",
"decode_symbolic_representation",
"decode_embodied_simulation",
"decode_ethical_reasoning",
"decode_proactive_behavior",
"decode_explainability_levels",
"decode_rl_integration",
"decode_fl_compatibility",
"decode_dp_features",
"decode_robustness_metrics",
"decode_calibration_score",
"decode_ood_detection"
],
"eos_token_id": [
128001,
128008,
128009
],
"fusion": {
"details": "structural_fusion_not_applied_by_script",
"layers_original": 28
},
"head_dim": 128,
"hidden_act": "silu",
"hidden_size": 3072,
"initializer_range": 0.02,
"instruction_tuning_status": "Conceptual - Designed/Processed for instruction following. Actual fine-tuning may be required depending on base model.",
"intermediate_size": 8192,
"is_instruct_model": true,
"is_trained": true,
"max_position_embeddings": 8000,
"mlp_bias": false,
"model_type": "llama",
"num_attention_heads": 24,
"num_hidden_layers": 28,
"num_key_value_heads": 8,
"pad_token_id": 128004,
"pretraining_tp": 1,
"processing_pipeline": [
"load_float32",
"add_special_tokens",
"resize_embeddings",
"dynamic_quantize_8bit",
"zero_biases",
"set_eval_mode",
"set_instruct_flag",
"update_config",
"add_custom_files",
"push_to_hub"
],
"rms_norm_eps": 1e-05,
"rope_scaling": {
"factor": 32.0,
"high_freq_factor": 4.0,
"low_freq_factor": 1.0,
"original_max_position_embeddings": 8192,
"rope_type": "llama3"
},
"rope_theta": 500000.0,
"safetensors": true,
"tensor_fusion": true,
"tensor_fusion_size": 3606776832,
"tie_word_embeddings": true,
"torch_dtype": "float32",
"training_notes": "Model has been processed from a pre-trained version. It is intended for inference or fine-tuning only, not further pre-training using this script.",
"transformers_version": "4.51.3",
"unsloth_version": "2025.2.15",
"use_cache": true,
"vocab_size": 128260
}
|