{ "cells": [ { "cell_type": "code", "source": [ "!pip install langgraph transformers accelerate bitsandbytes --quiet\n", "!pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 --quiet\n", "!pip install huggingface_hub --quiet\n", "print(\"✅ All packages installed successfully!\")" ], "metadata": { "colab": { "base_uri": "https://localhost:8080/" }, "id": "pHCUMUcpuUKP", "outputId": "d39b2ecb-8267-4c14-8583-94816c9d7cc8" }, "execution_count": 1, "outputs": [ { "output_type": "stream", "name": "stdout", "text": [ "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m43.7/43.7 kB\u001b[0m \u001b[31m3.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m143.7/143.7 kB\u001b[0m \u001b[31m10.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m72.9/72.9 MB\u001b[0m \u001b[31m11.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m43.8/43.8 kB\u001b[0m \u001b[31m1.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m50.1/50.1 kB\u001b[0m \u001b[31m4.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m363.4/363.4 MB\u001b[0m \u001b[31m4.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m13.8/13.8 MB\u001b[0m \u001b[31m74.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m24.6/24.6 MB\u001b[0m \u001b[31m57.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m883.7/883.7 kB\u001b[0m \u001b[31m46.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m664.8/664.8 MB\u001b[0m \u001b[31m2.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m211.5/211.5 MB\u001b[0m \u001b[31m6.9 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m56.3/56.3 MB\u001b[0m \u001b[31m18.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m127.9/127.9 MB\u001b[0m \u001b[31m7.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m207.5/207.5 MB\u001b[0m \u001b[31m5.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m21.1/21.1 MB\u001b[0m \u001b[31m33.9 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m216.5/216.5 kB\u001b[0m \u001b[31m14.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", "\u001b[?25h✅ All packages installed successfully!\n" ] } ] }, { "cell_type": "code", "source": [ "import torch\n", "print(f\"CUDA available: {torch.cuda.is_available()}\")\n", "if torch.cuda.is_available():\n", " print(f\"GPU device: {torch.cuda.get_device_name(0)}\")\n", " print(f\"CUDA version: {torch.version.cuda}\")\n", " print(f\"GPU memory: {torch.cuda.get_device_properties(0).total_memory / 1024**3:.2f} GB\")\n", "else:\n", " print(\"⚠️ No GPU detected. Please ensure you're using a GPU runtime.\")" ], "metadata": { "colab": { "base_uri": "https://localhost:8080/" }, "id": "AQcbO1EauYn6", "outputId": "6c19ea95-990a-49c5-be96-fc5794af0fb7" }, "execution_count": 2, "outputs": [ { "output_type": "stream", "name": "stdout", "text": [ "CUDA available: True\n", "GPU device: Tesla T4\n", "CUDA version: 12.4\n", "GPU memory: 14.74 GB\n" ] } ] }, { "cell_type": "code", "source": [ "from huggingface_hub import notebook_login\n", "print(\"Please login to Hugging Face to access Llama models:\")\n", "notebook_login()" ], "metadata": { "colab": { "base_uri": "https://localhost:8080/", "height": 34, "referenced_widgets": [ "6aa6b3eacc2e48008d91b19b550777b0", "82a109468484437589a237ad6aa6ccd0", "c8c6feab48974a6ebee87baf09e4fac9", "ef3dc0e417554f82b4ed5d035fd6db3e", "8ba1e396fa98444a8f9a4d7c934a01ab", "4a7a2ec05bb84bf08226b02f6f4ba7e1", "e0506e2bf8344391b1e9ecf74bed1c78", "2c209a33661b44429357853d4559a1d9", "27153a4694b4428384f2aafdbcff12ae", "4a385f45e98e4764ae09e0ad5fdce877", "a6f70a16108f4f3290622c192447c279", "24c43fa26cfa4d9eadd6d93a017c1b68", "bfcad1bf7e7f4080b599faddec0f1a44", "b269fe14fee24cb48bf3e384ed6adccc", "db69dea8814b4665b2aafa13f6488208", "023bb04d332742dc836360b4f8588fa0", "10e91aebff1241e68ee603efac3cf801", "eab4314829d64972b74f2e85b9b1cfbd", "ea16f9eb434445209243e647a8a702d4", "3caad989d7fe41cdbf57780bafe5c810" ] }, "id": "EyeFvFx8uZI4", "outputId": "19cdafbf-b9f5-405b-c1f0-49ae39f1477d" }, "execution_count": 6, "outputs": [ { "output_type": "stream", "name": "stdout", "text": [ "Please login to Hugging Face to access Llama models:\n" ] }, { "output_type": "display_data", "data": { "text/plain": [ "VBox(children=(HTML(value='