You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

πŸ¦… Falcon-7B Model Card (MarkAI Hosted Version)

Falcon Logo

Model Overview

Falcon-7B is a 7 billion parameter causal decoder-only model developed by Technology Innovation Institute (TII). This repository hosts the original model weights as part of MarkAI's model collection.

Technical Specifications

Architecture

Component Specification
Model Type Causal Decoder-only
Attention Mechanism Multi-Query + FlashAttention
Positional Embeddings Rotary Positional Embeddings
Normalization Single LayerNorm per block

Training Details

Parameter Value
Training Tokens 1,500B (1.5 trillion)
Training Compute 384 Γ— A100 40GB GPUs (P4d instances)
Training Time ~2 weeks
Precision bfloat16

Usage Examples

Text Generation

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "ibrahimlasfar/MarkAI",
    device_map="auto",
    torch_dtype=torch.bfloat16,
    trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained("ibrahimlasfar/MarkAI")

inputs = tokenizer(
    "The future of artificial intelligence",
    return_tensors="pt"
).to("cuda")
outputs = model.generate(
    **inputs,
    max_length=100,
    do_sample=True,
    top_k=10
)
print(tokenizer.decode(outputs[0]))
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for ibrahimlasfar/MarkAI

Base model

tiiuae/falcon-7b
Finetuned
(45)
this model

Dataset used to train ibrahimlasfar/MarkAI