Spaces:
Sleeping
Sleeping
File size: 1,397 Bytes
14e24b1 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 | ---
title: "Phi-3 Mini 128K Chat"
emoji: "π¬"
colorFrom: "blue"
colorTo: "purple"
sdk: "gradio"
python_version: "3.10"
app_file: "app.py"
suggested_hardware: "a10g-small"
suggested_storage: "medium"
short_description: "A demo of Phi-3-mini-128k-instruct model."
tags:
- phi-3
- microsoft
- chatbot
- long-context
- transformers
- gradio
models:
- microsoft/Phi-3-mini-128k-instruct
preload_from_hub:
- microsoft/Phi-3-mini-128k-instruct
disable_embedding: false
fullscreen: true
---
# Phi-3 Mini 128K Instruct Chat Demo
This Space demonstrates **Microsoft's [Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct)**, a powerful small language model with support for **up to 128,000 tokens of context**.
π§ Built with:
- π€ Transformers
- Gradio (for UI)
- Hugging Face Hub
π Features:
- Long-context understanding
- Instruction-tuned chat format
- Fast response generation
- Optimized for clarity and reasoning
π‘ Try asking it to:
- Summarize long texts
- Explain complex topics
- Write code
- Answer questions with context
> β³ **Note**: First load may take 2β3 minutes as the model initializes. Subsequent runs are faster thanks to `preload_from_hub`.
---
## How It Works
The model is preloaded during build time using `preload_from_hub`, ensuring minimal cold-start delay. The chat interface uses the official chat template:
|