File size: 1,793 Bytes
8b6a7b5 cce6826 e679c16 13561af e679c16 13561af e679c16 13561af e679c16 8b6a7b5 13561af e679c16 f510ceb e679c16 f510ceb e679c16 f510ceb e679c16 cce6826 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 |
---
language:
- en
tags:
- coding
- ui
- chat
- math
- factual
- agent
- multimodal
base_model:
- mistralai/Mixtral-8x7B-v0.1
- OpenBuddy/openbuddy-openllama-7b-v12-bf16
- HuggingFaceH4/mistral-7b-grok
- togethercomputer/RedPajama-INCITE-7B-Chat
datasets:
- hotboxxgenn/mix-openhermes-openorca-platypus-airoboros-chatalpaca-opencode
- microsoft/rStar-Coder
- ed001/ds-coder-instruct-v1
- bigcode/starcoderdata
- bigcode/starcoder2data-extras
- codeparrot/self-instruct-starcoder
- mrtoy/mobile-ui-design
- YashJain/UI-Elements-Detection-Dataset
- tecky-tech/Tecky-UI-Elements-VLM
- Tesslate/UIGEN-T2
- FineWeb
- OpenWebMath
- UltraChat
- WizardCoderData
library_name: transformers
---
# ๐ BerryAI
**Author:** [@hotboxxgenn](https://huggingface.co/hotboxxgenn)
**Version:** 1.1
**Type:** Conversational + Coding + UI + Math + Factual Model
**Base:** Mixtral 8x7B, OpenBuddy, Mistral-Grok, RedPajama
---
## โจ Overview
BerryAI is a **multi-skill LLM** designed to perform:
- ๐ป **Coding** โ Python, JS, React, Tailwind, multi-step reasoning
- ๐จ **UI generation** โ generate clean, responsive interfaces
- ๐ฌ **Conversational chat** โ helpful, creative, engaging tone
- ๐งฎ **Math reasoning** โ step-by-step calculations
- ๐ **Factual grounding** โ reduced hallucination, improved accuracy
---
## ๐ Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("hotboxxgenn/BerryAI")
model = AutoModelForCausalLM.from_pretrained("hotboxxgenn/BerryAI")
prompt = "Generate a responsive React login form with Tailwind CSS."
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |