--- license: apache-2.0 base_model: qwen3 tags: - affine - qwen3 - causal-lm - reasoning library_name: transformers pipeline_tag: text-generation --- # Model Card ## Description A Qwen3-based language model (~7B parameters) optimized for the Affine network. Features a 40K token context window, 36 transformer layers, and efficient grouped query attention (GQA) architecture. Designed for high-performance reasoning, code generation, and agentic AI applications. ## What is this used for? - **Complex Reasoning**: Multi-step problem solving and logical deduction - **Code Generation**: Python, JavaScript, and other programming languages - **Agentic AI**: Tool-using agents and autonomous systems - **Long-Context Tasks**: Document analysis and research - **Affine Network**: Competitive reasoning model for decentralized evaluation ## Quick Start ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model_name = "your-username/your-model-name" tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype=torch.bfloat16, device_map="auto", trust_remote_code=True ) prompt = "Explain quantum computing." inputs = tokenizer(prompt, return_tensors="pt").to(model.device) outputs = model.generate(**inputs, max_new_tokens=512) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Model Details - **Architecture**: Qwen3ForCausalLM - **Parameters**: ~7B - **Context Length**: 40,960 tokens - **Layers**: 36 - **Precision**: bfloat16 ## License Apache 2.0