Dataset Viewer
Auto-converted to Parquet Duplicate
entries
list
last_updated
null
version
string
schema
dict
competition
dict
[]
null
1.0
{ "name": "Participant name", "team": "Team name (optional)", "perplexity": "Model perplexity - lower is better", "loss": "Final validation cross-entropy loss", "tokens_per_sec": "Training throughput", "timestamp": "ISO 8601 submission timestamp", "model_config": "Optional model configuration details" }
{ "name": "Transformer Hackathon", "training_time_minutes": 45, "dataset": "TinyStories", "github": "https://github.com/abhishekadile/Transformer_Repo-" }

πŸ† Transformer Hackathon Leaderboard

Build Your Own GPT β€’ Train β€’ Compete β€’ Win!

GitHub License


πŸ“Š Current Standings

Rank Name Team Perplexity ⬇️ Loss Tokens/sec
πŸ₯‡ Be the first! - - - -

Lower perplexity = better model performance


🎯 Competition Categories

Category Metric Prize
πŸ† Best Performance Lowest Perplexity 1st Place
⚑ Most Efficient Highest Tokens/sec Speed Award
πŸ“ Best Generation Text Quality Creativity Award

πŸš€ How to Participate

# Clone the repository
git clone https://github.com/abhishekadile/Transformer_Repo-
cd Transformer_Repo-

# Install dependencies
pip install -r requirements.txt

# Run the hackathon pipeline
python run_hackathon.py

The script will:

  1. Download the TinyStories dataset
  2. Train your model for 45 minutes
  3. Evaluate performance
  4. Submit to this leaderboard!

πŸ“ˆ Leaderboard Schema

Field Type Description
name string Participant name
team string Team name
perplexity float Model perplexity (lower = better)
loss float Final validation loss
tokens_per_sec float Training speed
timestamp string Submission time (UTC)

πŸ”§ Optimization Tips

  • πŸš€ Enable mixed precision: --use-amp
  • πŸ“¦ Increase batch size if GPU allows
  • πŸ“ˆ Try different learning rates
  • 🧠 Experiment with model size in config.py

πŸ“š Resources


Good luck and have fun! πŸŽ‰

Downloads last month
20

Paper for abhisu30/transformer-hackathon-leaderboard