Phi-3.5 Mini Instruct โ€“ Uncensored Edition

This repository provides the Phi-3.5 Mini Instruct โ€“ Uncensored model โ€” a compact, instruction-fine-tuned language model variant optimized for flexible local usage and reduced behavioral filtering.


Model Overview

Model Name: Phi-3.5 Mini Instruct โ€“ Uncensored
Base Architecture: Microsoft Phi-3.5 Mini
Parameter Size: ~3.8B parameters
Maintainer / Publisher: SicariusSicariiStuff
License: Inherits the licensing terms of the original Phi-3.5 release
Primary Focus: Lightweight, instruction-optimized assistant with relaxed response constraints for local and research environments


About This Model

Phi-3.5 Mini Instruct โ€“ Uncensored is derived from Microsoftโ€™s Phi-3.5 Mini architecture and further adapted to reduce conversational restrictions while maintaining instruction-following performance.

This version is intended for users who:

  • Prefer greater control over response boundaries
  • Conduct experimentation with alignment styles
  • Run models locally on consumer hardware
  • Require a compact yet capable instruction model

The emphasis is on responsiveness, clarity, and adaptability in extended conversations.


Conversation Format

The model follows a structured instruction-style prompt format compatible with common chat-based runtimes.

Example structure:

<|system|>
System instructions or behavior guidelines
<|end|>
<|user|>
User input prompt
<|end|>
<|assistant|>
Model response

Depending on the runtime (Transformers, LM Studio, Ollama, etc.), the exact formatting may be handled automatically.


Core Characteristics

  • Instruction-tuned for dialogue-based interaction
  • Reduced filtering compared to standard releases
  • Compact parameter size suitable for mid-range GPUs
  • Capable of multi-step reasoning and contextual continuity
  • Strong performance relative to its size class
  • Works well in both interactive chat and structured prompting

Recommended Use Cases

Private AI assistant deployments โ€” local desktop or server usage
Development & scripting help โ€” explanations, snippets, structured answers
Prompt experimentation โ€” alignment and response-style research
Creative writing & roleplay systems โ€” dynamic, adaptive interactions
Edge-device experimentation โ€” lightweight inference setups


Deployment Compatibility

Compatible with:

  • Hugging Face Transformers
  • GGUF / llama.cpp-based runtimes (if converted)
  • LM Studio
  • Text-generation-webui
  • Other inference frameworks supporting Phi architecture

Performance will vary depending on quantization level and hardware configuration.


Notes on Behavior

As an uncensored variant, this model may produce outputs that are less restricted than official releases. It is intended for controlled, private, or research-oriented environments where users manage deployment context responsibly.


Acknowledgements

Credit to:

  • Microsoft for the original Phi-3.5 architecture
  • Open-source contributors supporting quantization and deployment tooling
  • Community members testing and refining local instruction models
Downloads last month
204
GGUF
Model size
4B params
Architecture
phi3
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Andycurrent/Phi-3.5-mini-instruct_Uncensored_GGUF

Collection including Andycurrent/Phi-3.5-mini-instruct_Uncensored_GGUF