{ "cells": [ { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ” Hugging Face Authentication for Google Colab\n", "try:\n", " from google.colab import userdata\n", " import os\n", " hf_token = userdata.get('HF_TOKEN')\n", " os.environ['HUGGINGFACE_HUB_TOKEN'] = hf_token\n", " print('โœ… HF token loaded from Colab secrets')\n", "except ImportError:\n", " print('โš ๏ธ Not running in Colab, skipping token setup')\n", "except Exception as e:\n", " print(f'โš ๏ธ Could not load HF_TOKEN from Colab secrets: {e}')\n", " print('๐Ÿ’ก Add HF_TOKEN to Colab secrets: Secrets tab โ†’ Add new secret โ†’ Name: HF_TOKEN')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”ง Install compatible versions for stable training\n", "!pip install -q transformers>=4.36.0 tokenizers>=0.15.0\n", "!pip install -q peft>=0.8.0 datasets>=2.16.0 bitsandbytes>=0.42.0 accelerate>=0.26.0 huggingface_hub trl\n", "import os; os.environ['TOKENIZERS_PARALLELISM'] = 'false'\n", "print('โœ… Compatible HF stack installed')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ›ก๏ธ Safe loading functions to avoid tokenizer and import errors\n", "from transformers import AutoTokenizer, AutoModelForCausalLM\n", "\n", "def safe_load_tokenizer(model_name, **kwargs):\n", " \"\"\"Load tokenizer with safe defaults\"\"\"\n", " kwargs.setdefault('use_fast', False)\n", " kwargs.setdefault('trust_remote_code', False)\n", " return AutoTokenizer.from_pretrained(model_name, **kwargs)\n", "\n", "def safe_load_model(model_name, **kwargs):\n", " \"\"\"Load model with safe defaults\"\"\"\n", " kwargs.setdefault('trust_remote_code', False)\n", " return AutoModelForCausalLM.from_pretrained(model_name, **kwargs)\n", "\n", "print('โœ… Safe loading functions ready')\n", "print('๐Ÿ’ก Use: tokenizer = safe_load_tokenizer(MODEL_NAME)')\n", "print('๐Ÿ’ก Use: model = safe_load_model(MODEL_NAME, quantization_config=bnb_config, device_map=\"auto\")')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ” Hugging Face Authentication for Google Colab\n", "try:\n", " from google.colab import userdata\n", " import os\n", " hf_token = userdata.get('HF_TOKEN')\n", " os.environ['HUGGINGFACE_HUB_TOKEN'] = hf_token\n", " print('โœ… HF token loaded from Colab secrets')\n", "except ImportError:\n", " print('โš ๏ธ Not running in Colab, skipping token setup')\n", "except Exception as e:\n", " print(f'โš ๏ธ Could not load HF_TOKEN from Colab secrets: {e}')\n", " print('๐Ÿ’ก Add HF_TOKEN to Colab secrets: Secrets tab โ†’ Add new secret โ†’ Name: HF_TOKEN')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”ง Install compatible versions for stable training\n", "!pip install -q transformers>=4.36.0 tokenizers>=0.15.0\n", "!pip install -q peft>=0.8.0 datasets>=2.16.0 bitsandbytes>=0.42.0 accelerate>=0.26.0 huggingface_hub trl\n", "import os; os.environ['TOKENIZERS_PARALLELISM'] = 'false'\n", "print('โœ… Compatible HF stack installed')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”„ Clear any previous patches and restart imports\n", "import importlib\n", "import sys\n", "\n", "# Clear transformers from cache if it exists\n", "if 'transformers' in sys.modules:\n", " del sys.modules['transformers']\n", " print('๐Ÿงน Cleared transformers from module cache')\n", "\n", "# Fresh import\n", "from transformers import AutoTokenizer, AutoModelForCausalLM\n", "print('โœ… Fresh transformers import - no patches applied')\n", "print('๐Ÿ’ก Use explicit parameters: AutoTokenizer.from_pretrained(model, use_fast=False, trust_remote_code=False)')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ” Hugging Face Authentication for Google Colab\n", "try:\n", " from google.colab import userdata\n", " import os\n", " hf_token = userdata.get('HF_TOKEN')\n", " os.environ['HUGGINGFACE_HUB_TOKEN'] = hf_token\n", " print('โœ… HF token loaded from Colab secrets')\n", "except ImportError:\n", " print('โš ๏ธ Not running in Colab, skipping token setup')\n", "except Exception as e:\n", " print(f'โš ๏ธ Could not load HF_TOKEN from Colab secrets: {e}')\n", " print('๐Ÿ’ก Add HF_TOKEN to Colab secrets: Secrets tab โ†’ Add new secret โ†’ Name: HF_TOKEN')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”ง Install compatible versions for stable training\n", "!pip install -q transformers>=4.36.0 tokenizers>=0.15.0\n", "!pip install -q peft>=0.8.0 datasets>=2.16.0 bitsandbytes>=0.42.0 accelerate>=0.26.0 huggingface_hub trl\n", "import os; os.environ['TOKENIZERS_PARALLELISM'] = 'false'\n", "print('โœ… Compatible HF stack installed')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿฉน Force safe defaults to avoid fast-tokenizer and remote code import issues\n", "import transformers\n", "from transformers import AutoTokenizer, AutoModelForCausalLM\n", "\n", "# Store original methods\n", "_orig_tok_from_pretrained = AutoTokenizer.from_pretrained.__func__\n", "_orig_model_from_pretrained = AutoModelForCausalLM.from_pretrained.__func__\n", "\n", "# Create safe wrapper functions\n", "def safe_tokenizer_from_pretrained(cls, *args, **kwargs):\n", " kwargs.setdefault('use_fast', False)\n", " kwargs.setdefault('trust_remote_code', False)\n", " return _orig_tok_from_pretrained(cls, *args, **kwargs)\n", "\n", "def safe_model_from_pretrained(cls, *args, **kwargs):\n", " kwargs.setdefault('trust_remote_code', False)\n", " return _orig_model_from_pretrained(cls, *args, **kwargs)\n", "\n", "# Apply patches\n", "AutoTokenizer.from_pretrained = classmethod(safe_tokenizer_from_pretrained)\n", "AutoModelForCausalLM.from_pretrained = classmethod(safe_model_from_pretrained)\n", "print('โœ… Patched: AutoTokenizer(use_fast=False, trust_remote_code=False) and AutoModel(trust_remote_code=False) by default')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ” Hugging Face Authentication for Google Colab\n", "try:\n", " from google.colab import userdata\n", " import os\n", " hf_token = userdata.get('HF_TOKEN')\n", " os.environ['HUGGINGFACE_HUB_TOKEN'] = hf_token\n", " print('โœ… HF token loaded from Colab secrets')\n", "except ImportError:\n", " print('โš ๏ธ Not running in Colab, skipping token setup')\n", "except Exception as e:\n", " print(f'โš ๏ธ Could not load HF_TOKEN from Colab secrets: {e}')\n", " print('๐Ÿ’ก Add HF_TOKEN to Colab secrets: Secrets tab โ†’ Add new secret โ†’ Name: HF_TOKEN')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”ง Install pinned versions for stable training\n", "!pip install -q transformers==4.46.2 tokenizers==0.20.1\n", "!pip install -q peft==0.14.0 datasets==2.20.0 bitsandbytes==0.43.3 accelerate==0.34.2 huggingface_hub==0.24.6 trl==0.11.4\n", "import os; os.environ['TOKENIZERS_PARALLELISM'] = 'false'\n", "print('โœ… Pinned HF stack installed')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿฉน Force safe defaults to avoid fast-tokenizer and remote code import issues\n", "from transformers import AutoTokenizer as _AutoTokenizer, AutoModelForCausalLM as _AutoModelForCausalLM\n", "_orig_tok_from_pretrained = _AutoTokenizer.from_pretrained\n", "def _patched_tok_from_pretrained(*args, **kwargs):\n", " kwargs.setdefault('use_fast', False)\n", " kwargs.setdefault('trust_remote_code', False)\n", " return _orig_tok_from_pretrained(*args, **kwargs)\n", "_AutoTokenizer.from_pretrained = staticmethod(_patched_tok_from_pretrained)\n", "\n", "_orig_model_from_pretrained = _AutoModelForCausalLM.from_pretrained\n", "def _patched_model_from_pretrained(*args, **kwargs):\n", " kwargs.setdefault('trust_remote_code', False)\n", " return _orig_model_from_pretrained(*args, **kwargs)\n", "_AutoModelForCausalLM.from_pretrained = staticmethod(_patched_model_from_pretrained)\n", "print('โœ… Patched: AutoTokenizer(use_fast=False, trust_remote_code=False) and AutoModel(trust_remote_code=False) by default')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”ง Install pinned versions for stable training\n", "!pip install -q transformers==4.46.2 tokenizers==0.20.1\n", "!pip install -q peft==0.14.0 datasets==2.20.0 bitsandbytes==0.43.3 accelerate==0.34.2 huggingface_hub==0.24.6 trl==0.11.4\n", "import os; os.environ['TOKENIZERS_PARALLELISM'] = 'false'\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿฉน Force safe defaults to avoid fast-tokenizer and remote code import issues\n", "from transformers import AutoTokenizer as _AutoTokenizer, AutoModelForCausalLM as _AutoModelForCausalLM\n", "_orig_tok_from_pretrained = _AutoTokenizer.from_pretrained\n", "def _patched_tok_from_pretrained(*args, **kwargs):\n", " kwargs.setdefault('use_fast', False)\n", " kwargs.setdefault('trust_remote_code', False)\n", " return _orig_tok_from_pretrained(*args, **kwargs)\n", "_AutoTokenizer.from_pretrained = staticmethod(_patched_tok_from_pretrained)\n", "\n", "_orig_model_from_pretrained = _AutoModelForCausalLM.from_pretrained\n", "def _patched_model_from_pretrained(*args, **kwargs):\n", " kwargs.setdefault('trust_remote_code', False)\n", " return _orig_model_from_pretrained(*args, **kwargs)\n", "_AutoModelForCausalLM.from_pretrained = staticmethod(_patched_model_from_pretrained)\n", "print('โœ… Patched: AutoTokenizer(use_fast=False, trust_remote_code=False) and AutoModel(trust_remote_code=False) by default')\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿ”ง Install pinned versions for stable training\n", "!pip install -q transformers==4.46.2 tokenizers==0.20.1\n", "!pip install -q peft==0.14.0 datasets==2.20.0 bitsandbytes==0.43.3 accelerate==0.34.2 huggingface_hub==0.24.6 trl==0.11.4\n", "import os; os.environ['TOKENIZERS_PARALLELISM'] = 'false'\n" ] }, { "cell_type": "code", "metadata": {}, "execution_count": null, "outputs": [], "source": [ "# ๐Ÿฉน Force slow tokenizer by default to avoid PyPreTokenizerTypeWrapper errors\n", "from transformers import AutoTokenizer as _AutoTokenizer\n", "_orig_from_pretrained = _AutoTokenizer.from_pretrained\n", "def _patched_from_pretrained(*args, **kwargs):\n", " kwargs.setdefault('use_fast', False)\n", " return _orig_from_pretrained(*args, **kwargs)\n", "_AutoTokenizer.from_pretrained = staticmethod(_patched_from_pretrained)\n", "print('โœ… Patched AutoTokenizer.from_pretrained to default use_fast=False')\n" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "# ๐ŸŒŸ CELESTIAL MISTRAL 7B TRAINING\n", "## Train Your Own Mistral 7B Model for CELESTIAL AI\n", "\n", "This notebook properly trains Mistral 7B v0.3 with:\n", "- 150 production-quality conversations\n", "- LoRA fine-tuning for efficiency\n", "- Proper chat formatting for Mistral\n", "- No logging issues" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿ“ฆ INSTALL REQUIRED PACKAGES FOR MISTRAL 7B\n", "!pip install -q transformers==4.36.0 datasets accelerate peft bitsandbytes huggingface_hub trl\n", "\n", "# Disable all logging to prevent issues\n", "import os\n", "import warnings\n", "os.environ[\"WANDB_DISABLED\"] = \"true\"\n", "os.environ[\"WANDB_MODE\"] = \"disabled\"\n", "os.environ[\"TOKENIZERS_PARALLELISM\"] = \"false\"\n", "warnings.filterwarnings('ignore')\n", "\n", "print('โœ… Packages installed for Mistral 7B training!')\n", "print('๐Ÿšซ All logging disabled to prevent errors')" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿ”‘ HUGGINGFACE AUTHENTICATION\n", "from huggingface_hub import notebook_login\n", "\n", "print('๐Ÿ” Authenticating with HuggingFace for Mistral access...')\n", "try:\n", " notebook_login()\n", " print('โœ… Authentication successful!')\n", "except Exception as e:\n", " print(f'โš ๏ธ Authentication failed: {e}')\n", " print('Please set your HF token manually if needed')" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿ“Š LOAD CELESTIAL DATASET\n", "from datasets import load_dataset\n", "\n", "DATASET_REPO = 'dp1812/celestial-comprehensive-spiritual-ai'\n", "\n", "print('๐Ÿ“Š Loading CELESTIAL dataset for Mistral training...')\n", "try:\n", " dataset = load_dataset(DATASET_REPO, data_files='celestial_complete_production_dataset.jsonl', split='train')\n", " print(f'โœ… Dataset loaded: {len(dataset)} conversations')\n", " print('๐ŸŽฏ 100 numerology + 50 Krishna divine guidance')\n", "except Exception as e:\n", " print(f'โŒ Dataset loading failed: {e}')\n", " # Fallback\n", " try:\n", " dataset = load_dataset(DATASET_REPO, split='train')\n", " print(f'โœ… Fallback dataset loaded: {len(dataset)} conversations')\n", " except Exception as e2:\n", " print(f'โŒ All dataset loading failed: {e2}')\n", " raise\n", "\n", "# Show sample\n", "print('\\n๐Ÿ“ Sample conversation:')\n", "sample = dataset[0]\n", "print(f\"User: {sample['messages'][1]['content'][:60]}...\")\n", "print(f\"Assistant: {sample['messages'][2]['content'][:60]}...\")" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿค– LOAD MISTRAL 7B MODEL AND TOKENIZER\n", "from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig\n", "import torch\n", "\n", "MODEL_NAME = 'mistralai/Mistral-7B-v0.3'\n", "\n", "print('๐Ÿค– Loading Mistral 7B v0.3 model and tokenizer...')\n", "\n", "# Load tokenizer with proper settings\n", "tokenizer = AutoTokenizer.from_pretrained(\n", " MODEL_NAME,\n", " trust_remote_code=True,\n", " padding_side='right'\n", ")\n", "\n", "# Add pad token if missing\n", "if tokenizer.pad_token is None:\n", " tokenizer.pad_token = tokenizer.eos_token\n", " tokenizer.pad_token_id = tokenizer.eos_token_id\n", "\n", "# Quantization config for efficient training\n", "bnb_config = BitsAndBytesConfig(\n", " load_in_4bit=True,\n", " bnb_4bit_quant_type=\"nf4\",\n", " bnb_4bit_compute_dtype=torch.float16,\n", " bnb_4bit_use_double_quant=True\n", ")\n", "\n", "# Load Mistral 7B model\n", "model = AutoModelForCausalLM.from_pretrained(\n", " MODEL_NAME,\n", " quantization_config=bnb_config,\n", " device_map=\"auto\",\n", " trust_remote_code=True,\n", " torch_dtype=torch.float16\n", ")\n", "\n", "print('โœ… Mistral 7B model and tokenizer loaded successfully!')\n", "print(f'๐Ÿ” Model: {MODEL_NAME}')\n", "print(f'๐Ÿ” Tokenizer vocab size: {len(tokenizer)}')\n", "print(f'๐Ÿ” Model device: {model.device}')" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿ”ง SETUP LORA FOR MISTRAL 7B\n", "from peft import LoraConfig, get_peft_model, TaskType, prepare_model_for_kbit_training\n", "\n", "print('๐Ÿ”ง Setting up LoRA for Mistral 7B training...')\n", "\n", "# Prepare model for k-bit training\n", "model = prepare_model_for_kbit_training(model)\n", "\n", "# Mistral 7B specific target modules\n", "target_modules = [\n", " \"q_proj\",\n", " \"k_proj\", \n", " \"v_proj\",\n", " \"o_proj\",\n", " \"gate_proj\",\n", " \"up_proj\",\n", " \"down_proj\",\n", " \"lm_head\"\n", "]\n", "\n", "print(f'๐ŸŽฏ Target modules for Mistral: {target_modules}')\n", "\n", "# Create LoRA config optimized for Mistral\n", "lora_config = LoraConfig(\n", " r=64, # Higher rank for better performance\n", " lora_alpha=16,\n", " target_modules=target_modules,\n", " lora_dropout=0.1,\n", " bias=\"none\",\n", " task_type=TaskType.CAUSAL_LM,\n", ")\n", "\n", "# Apply LoRA to Mistral\n", "try:\n", " model = get_peft_model(model, lora_config)\n", " model.print_trainable_parameters()\n", " print('โœ… LoRA adapters attached to Mistral 7B!')\n", "except Exception as e:\n", " print(f'โŒ LoRA setup failed: {e}')\n", " raise\n", "\n", "print('๐ŸŽฏ Mistral 7B ready for CELESTIAL training!')" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿ“ FORMAT DATA FOR MISTRAL CHAT TRAINING\n", "def format_for_mistral_chat(example):\n", " \"\"\"Format conversation for Mistral chat training\"\"\"\n", " messages = example['messages']\n", " \n", " # Extract messages\n", " system_msg = messages[0]['content']\n", " user_msg = messages[1]['content']\n", " assistant_msg = messages[2]['content']\n", " \n", " # Mistral chat format\n", " formatted = f\"[INST] {system_msg}\\n\\nUser: {user_msg} [/INST] {assistant_msg}\"\n", " \n", " # Tokenize\n", " tokens = tokenizer(\n", " formatted,\n", " truncation=True,\n", " padding=False,\n", " max_length=2048, # Mistral context length\n", " return_tensors=None\n", " )\n", " \n", " # Set labels (same as input_ids for causal LM)\n", " tokens['labels'] = tokens['input_ids'].copy()\n", " \n", " return tokens\n", "\n", "print('๐Ÿ“ Formatting data for Mistral chat training...')\n", "formatted_dataset = dataset.map(\n", " format_for_mistral_chat,\n", " remove_columns=dataset.column_names,\n", " desc=\"Formatting for Mistral\"\n", ")\n", "\n", "print(f'โœ… Formatted {len(formatted_dataset)} conversations for Mistral')\n", "print('๐ŸŽฏ Using proper Mistral chat format with [INST] tags')" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿš€ MISTRAL TRAINING CONFIGURATION\n", "from transformers import TrainingArguments, Trainer, DataCollatorForLanguageModeling\n", "\n", "print('๐Ÿš€ Setting up Mistral 7B training configuration...')\n", "\n", "# Training arguments optimized for Mistral 7B\n", "training_args = TrainingArguments(\n", " output_dir='./celestial-mistral-7b-results',\n", " num_train_epochs=3,\n", " per_device_train_batch_size=1,\n", " gradient_accumulation_steps=16, # Effective batch size of 16\n", " warmup_steps=50,\n", " learning_rate=2e-4, # Higher LR for LoRA\n", " fp16=True,\n", " logging_steps=10,\n", " save_steps=100,\n", " eval_strategy='no',\n", " save_strategy='steps',\n", " load_best_model_at_end=False,\n", " report_to=[], # No external logging\n", " remove_unused_columns=False,\n", " dataloader_drop_last=True,\n", " group_by_length=True, # Efficient batching\n", " ddp_find_unused_parameters=False\n", ")\n", "\n", "# Data collator for Mistral\n", "data_collator = DataCollatorForLanguageModeling(\n", " tokenizer=tokenizer,\n", " mlm=False,\n", " pad_to_multiple_of=8\n", ")\n", "\n", "# Create Mistral trainer\n", "trainer = Trainer(\n", " model=model,\n", " args=training_args,\n", " train_dataset=formatted_dataset,\n", " tokenizer=tokenizer,\n", " data_collator=data_collator\n", ")\n", "\n", "print('โœ… Mistral 7B training configuration ready!')\n", "print('๐ŸŽฏ Optimized for CELESTIAL AI with LoRA fine-tuning')\n", "print('โฑ๏ธ Expected training time: 30-45 minutes')" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿƒโ€โ™‚๏ธ START MISTRAL 7B TRAINING\n", "print('๐Ÿƒโ€โ™‚๏ธ Starting CELESTIAL Mistral 7B training...')\n", "print('โฑ๏ธ Expected time: 30-45 minutes')\n", "print('๐ŸŽฏ Training Mistral 7B v0.3 on CELESTIAL conversations')\n", "print('๐Ÿ’Ž 150 production-quality conversations')\n", "print('\\n๐Ÿš€ Mistral training begins now...')\n", "\n", "try:\n", " # Start Mistral training\n", " trainer.train()\n", " \n", " print('\\n๐ŸŽ‰ MISTRAL 7B TRAINING COMPLETED SUCCESSFULLY!')\n", " print('โœ… CELESTIAL Mistral 7B is now trained!')\n", " print('๐ŸŒŸ Ready for testing and deployment!')\n", " \n", "except Exception as e:\n", " print(f'โŒ Mistral training failed: {e}')\n", " print('๐Ÿ”ง Please check the error and try again')\n", " raise" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# ๐Ÿงช TEST TRAINED MISTRAL 7B\n", "print('๐Ÿงช Testing the trained CELESTIAL Mistral 7B...')\n", "\n", "model.eval()\n", "\n", "test_prompts = [\n", " \"[INST] You are CELESTIAL AI, an expert numerologist. Provide detailed analysis.\\n\\nUser: Tell me about number 7 in Chaldean numerology. [/INST]\",\n", " \"[INST] You are Shree Krishna providing divine guidance.\\n\\nUser: Krishna, I need guidance about my career path. [/INST]\",\n", " \"[INST] You are CELESTIAL AI providing numerology analysis.\\n\\nUser: Calculate my numerology for name 'John Smith' born 15/08/1990. [/INST]\"\n", "]\n", "\n", "for i, prompt in enumerate(test_prompts, 1):\n", " print(f'\\n๐Ÿ” Test {i}: Mistral 7B Response')\n", " \n", " try:\n", " inputs = tokenizer(prompt, return_tensors=\"pt\").to(model.device)\n", " \n", " with torch.no_grad():\n", " outputs = model.generate(\n", " **inputs,\n", " max_new_tokens=300,\n", " temperature=0.7,\n", " do_sample=True,\n", " pad_token_id=tokenizer.pad_token_id,\n", " eos_token_id=tokenizer.eos_token_id\n", " )\n", " \n", " response = tokenizer.decode(outputs[0], skip_special_tokens=True)\n", " generated = response[len(prompt):].strip()\n", " \n", " print(f'๐Ÿค– Mistral Response: {generated[:250]}...')\n", " \n", " # Quality check\n", " if len(generated) > 50 and 'number' in generated.lower() or 'krishna' in generated.lower():\n", " print('โœ… Response quality: EXCELLENT')\n", " else:\n", " print('โš ๏ธ Response quality: NEEDS IMPROVEMENT')\n", " \n", " except Exception as e:\n", " print(f'โŒ Test {i} failed: {e}')\n", "\n", "print('\\n๐ŸŽ‰ CELESTIAL MISTRAL 7B TRAINING COMPLETE!')\n", "print('โœ… Your own trained Mistral 7B model is ready!')\n", "print('๐ŸŒŸ No external API dependencies - fully yours!')\n", "print('\\n๐Ÿš€ Next Steps:')\n", "print(' โ€ข Save the trained model to HuggingFace')\n", "print(' โ€ข Integrate with CELESTIAL platform')\n", "print(' โ€ข Expand training data for more features')\n", "print(' โ€ข Deploy to production environment')" ] } ], "metadata": { "kernelspec": { "display_name": "Python 3", "language": "python", "name": "python3" }, "language_info": { "codemirror_mode": { "name": "ipython", "version": 3 }, "file_extension": ".py", "name": "python", "nbconvert_exporter": "python", "pygments_lexer": "ipython3", "version": "3.8.5" } }, "nbformat": 4, "nbformat_minor": 4 }