File size: 1,397 Bytes
14e24b1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
---
title: "Phi-3 Mini 128K Chat"
emoji: "πŸ’¬"
colorFrom: "blue"
colorTo: "purple"
sdk: "gradio"
python_version: "3.10"
app_file: "app.py"
suggested_hardware: "a10g-small"
suggested_storage: "medium"
short_description: "A demo of Phi-3-mini-128k-instruct model."
tags:
  - phi-3
  - microsoft
  - chatbot
  - long-context
  - transformers
  - gradio
models:
  - microsoft/Phi-3-mini-128k-instruct
preload_from_hub:
  - microsoft/Phi-3-mini-128k-instruct
disable_embedding: false
fullscreen: true
---
# Phi-3 Mini 128K Instruct Chat Demo

This Space demonstrates **Microsoft's [Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct)**, a powerful small language model with support for **up to 128,000 tokens of context**.

πŸ”§ Built with:
- πŸ€— Transformers
- Gradio (for UI)
- Hugging Face Hub

πŸš€ Features:
- Long-context understanding
- Instruction-tuned chat format
- Fast response generation
- Optimized for clarity and reasoning

πŸ’‘ Try asking it to:
- Summarize long texts
- Explain complex topics
- Write code
- Answer questions with context

> ⏳ **Note**: First load may take 2–3 minutes as the model initializes. Subsequent runs are faster thanks to `preload_from_hub`.

---

## How It Works

The model is preloaded during build time using `preload_from_hub`, ensuring minimal cold-start delay. The chat interface uses the official chat template: