repo_code_snippets / source_snippets /SuperAlignment.jsonl
PeytonT's picture
Upload folder using huggingface_hub
d67f29d verified
{"repo_id":"SuperAlignment","entity_id":"py:train_example","uri":"program://SuperAlignment/module/train_example#L1-L60","kind":"module","name":"train_example","path":"train_example.py","language":"python","start_line":1,"end_line":60,"context_start_line":1,"context_end_line":60,"code":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision.models import resnet18, resnet50 # Example models\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True)\nstudent_model = resnet50(pretrained=True)\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss\n\n\n# Example Training Loop Skeleton\ndef train_model(student_model, teacher_model, dataloader, optimizer, device):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n loss_fn = AdaptiveConfidenceDistillationLoss(temperature=2.0)\n\n for inputs, labels in dataloader:\n inputs, labels = inputs.to(device), labels.to(device)\n\n optimizer.zero_grad()\n\n with torch.no_grad():\n teacher_outputs = teacher_model(inputs)\n\n student_outputs = student_model(inputs)\n\n loss = loss_fn(student_outputs, teacher_outputs, labels)\n loss.backward()\n optimizer.step()\n\n\n# Example usage\n# Assuming `dataloader`, `optimizer`, and `device` are defined\n# train_model(student_model, teacher_model, dataloader, optimizer, device)","source_hash":"1d853ce261864b7b90ca11a013e5ff6feff4aa1e87a9d5d20878083564c9aba6","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train_example.AdaptiveConfidenceDistillationLoss","uri":"program://SuperAlignment/class/train_example.AdaptiveConfidenceDistillationLoss#L12-L33","kind":"class","name":"AdaptiveConfidenceDistillationLoss","path":"train_example.py","language":"python","start_line":12,"end_line":33,"context_start_line":1,"context_end_line":53,"code":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision.models import resnet18, resnet50 # Example models\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True)\nstudent_model = resnet50(pretrained=True)\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss\n\n\n# Example Training Loop Skeleton\ndef train_model(student_model, teacher_model, dataloader, optimizer, device):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n loss_fn = AdaptiveConfidenceDistillationLoss(temperature=2.0)\n\n for inputs, labels in dataloader:\n inputs, labels = inputs.to(device), labels.to(device)\n\n optimizer.zero_grad()\n\n with torch.no_grad():\n teacher_outputs = teacher_model(inputs)\n\n student_outputs = student_model(inputs)\n\n loss = loss_fn(student_outputs, teacher_outputs, labels)","source_hash":"1d853ce261864b7b90ca11a013e5ff6feff4aa1e87a9d5d20878083564c9aba6","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train_example.train_model","uri":"program://SuperAlignment/function/train_example.train_model#L37-L55","kind":"function","name":"train_model","path":"train_example.py","language":"python","start_line":37,"end_line":55,"context_start_line":17,"context_end_line":60,"code":"\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss\n\n\n# Example Training Loop Skeleton\ndef train_model(student_model, teacher_model, dataloader, optimizer, device):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n loss_fn = AdaptiveConfidenceDistillationLoss(temperature=2.0)\n\n for inputs, labels in dataloader:\n inputs, labels = inputs.to(device), labels.to(device)\n\n optimizer.zero_grad()\n\n with torch.no_grad():\n teacher_outputs = teacher_model(inputs)\n\n student_outputs = student_model(inputs)\n\n loss = loss_fn(student_outputs, teacher_outputs, labels)\n loss.backward()\n optimizer.step()\n\n\n# Example usage\n# Assuming `dataloader`, `optimizer`, and `device` are defined\n# train_model(student_model, teacher_model, dataloader, optimizer, device)","source_hash":"1d853ce261864b7b90ca11a013e5ff6feff4aa1e87a9d5d20878083564c9aba6","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train_example.__init__","uri":"program://SuperAlignment/function/train_example.__init__#L13-L16","kind":"function","name":"__init__","path":"train_example.py","language":"python","start_line":13,"end_line":16,"context_start_line":1,"context_end_line":36,"code":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision.models import resnet18, resnet50 # Example models\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True)\nstudent_model = resnet50(pretrained=True)\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss\n\n\n# Example Training Loop Skeleton","source_hash":"1d853ce261864b7b90ca11a013e5ff6feff4aa1e87a9d5d20878083564c9aba6","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train_example.forward","uri":"program://SuperAlignment/function/train_example.forward#L18-L33","kind":"function","name":"forward","path":"train_example.py","language":"python","start_line":18,"end_line":33,"context_start_line":1,"context_end_line":53,"code":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision.models import resnet18, resnet50 # Example models\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True)\nstudent_model = resnet50(pretrained=True)\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss\n\n\n# Example Training Loop Skeleton\ndef train_model(student_model, teacher_model, dataloader, optimizer, device):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n loss_fn = AdaptiveConfidenceDistillationLoss(temperature=2.0)\n\n for inputs, labels in dataloader:\n inputs, labels = inputs.to(device), labels.to(device)\n\n optimizer.zero_grad()\n\n with torch.no_grad():\n teacher_outputs = teacher_model(inputs)\n\n student_outputs = student_model(inputs)\n\n loss = loss_fn(student_outputs, teacher_outputs, labels)","source_hash":"1d853ce261864b7b90ca11a013e5ff6feff4aa1e87a9d5d20878083564c9aba6","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:AdaptiveConfidenceDistillation","uri":"program://SuperAlignment/module/AdaptiveConfidenceDistillation#L1-L27","kind":"module","name":"AdaptiveConfidenceDistillation","path":"AdaptiveConfidenceDistillation.py","language":"python","start_line":1,"end_line":27,"context_start_line":1,"context_end_line":27,"code":"import torch.nn as nn\nimport torch.nn.functional as F\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss","source_hash":"78f2354cb5ad0726bbf11cff4469be7a070132d3339921881ae5df2398b71f16","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:AdaptiveConfidenceDistillation.AdaptiveConfidenceDistillationLoss","uri":"program://SuperAlignment/class/AdaptiveConfidenceDistillation.AdaptiveConfidenceDistillationLoss#L6-L27","kind":"class","name":"AdaptiveConfidenceDistillationLoss","path":"AdaptiveConfidenceDistillation.py","language":"python","start_line":6,"end_line":27,"context_start_line":1,"context_end_line":27,"code":"import torch.nn as nn\nimport torch.nn.functional as F\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss","source_hash":"78f2354cb5ad0726bbf11cff4469be7a070132d3339921881ae5df2398b71f16","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:AdaptiveConfidenceDistillation.__init__","uri":"program://SuperAlignment/function/AdaptiveConfidenceDistillation.__init__#L7-L10","kind":"function","name":"__init__","path":"AdaptiveConfidenceDistillation.py","language":"python","start_line":7,"end_line":10,"context_start_line":1,"context_end_line":27,"code":"import torch.nn as nn\nimport torch.nn.functional as F\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss","source_hash":"78f2354cb5ad0726bbf11cff4469be7a070132d3339921881ae5df2398b71f16","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:AdaptiveConfidenceDistillation.forward","uri":"program://SuperAlignment/function/AdaptiveConfidenceDistillation.forward#L12-L27","kind":"function","name":"forward","path":"AdaptiveConfidenceDistillation.py","language":"python","start_line":12,"end_line":27,"context_start_line":1,"context_end_line":27,"code":"import torch.nn as nn\nimport torch.nn.functional as F\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n\n # Calculate the standard loss with hard targets\n hard_loss = self.ce_loss(student_logits, targets)\n\n # Adaptive weighting could be implemented here based on confidence\n # For simplicity, this example uses a fixed ratio\n return distillation_loss + hard_loss","source_hash":"78f2354cb5ad0726bbf11cff4469be7a070132d3339921881ae5df2398b71f16","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:test_train_example","uri":"program://SuperAlignment/module/test_train_example#L1-L42","kind":"module","name":"test_train_example","path":"test_train_example.py","language":"python","start_line":1,"end_line":42,"context_start_line":1,"context_end_line":42,"code":"import torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torchvision.models import resnet18, resnet50\nfrom torch.utils.data import DataLoader, TensorDataset\nimport numpy as np\nfrom train_example import train_model\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=False) # Using pretrained=False for simplicity\nstudent_model = resnet50(pretrained=False)\n\n# Define a simple synthetic dataset resembling CIFAR-100\n# CIFAR-100 has 32x32 images with 3 channels\n# We'll create random data with similar dimensions\nnum_samples = 1000\ninput_channels = 3\ninput_height = 32\ninput_width = 32\nnum_classes = 100\n\n# Generate random images and labels\nimages = torch.randn(num_samples, input_channels, input_height, input_width)\nlabels = torch.randint(0, num_classes, (num_samples,))\n\n# Create DataLoader for the synthetic dataset\ndataset = TensorDataset(images, labels)\ndataloader = DataLoader(dataset, batch_size=32, shuffle=True)\n\n# Define optimizer and device\noptimizer = optim.Adam(student_model.parameters(), lr=0.001)\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n\n# Define a unit test function\ndef test_training_loop():\n train_model(student_model, teacher_model, dataloader, optimizer, device)\n\n\n# Run the test function\ntest_training_loop()\nprint(\"test passed!\")","source_hash":"6e808ad91b17c4e12df91f6ed188bfe810f9d6cb3ef55c61df0055a57b0abc0b","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:test_train_example.test_training_loop","uri":"program://SuperAlignment/function/test_train_example.test_training_loop#L36-L37","kind":"function","name":"test_training_loop","path":"test_train_example.py","language":"python","start_line":36,"end_line":37,"context_start_line":16,"context_end_line":42,"code":"num_samples = 1000\ninput_channels = 3\ninput_height = 32\ninput_width = 32\nnum_classes = 100\n\n# Generate random images and labels\nimages = torch.randn(num_samples, input_channels, input_height, input_width)\nlabels = torch.randint(0, num_classes, (num_samples,))\n\n# Create DataLoader for the synthetic dataset\ndataset = TensorDataset(images, labels)\ndataloader = DataLoader(dataset, batch_size=32, shuffle=True)\n\n# Define optimizer and device\noptimizer = optim.Adam(student_model.parameters(), lr=0.001)\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n\n# Define a unit test function\ndef test_training_loop():\n train_model(student_model, teacher_model, dataloader, optimizer, device)\n\n\n# Run the test function\ntest_training_loop()\nprint(\"test passed!\")","source_hash":"6e808ad91b17c4e12df91f6ed188bfe810f9d6cb3ef55c61df0055a57b0abc0b","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:dataset","uri":"program://SuperAlignment/module/dataset#L1-L58","kind":"module","name":"dataset","path":"dataset.py","language":"python","start_line":1,"end_line":58,"context_start_line":1,"context_end_line":58,"code":"import torch\nfrom torch.utils.data import Dataset\nimport os\nfrom PIL import Image\nfrom generate_labels import generate_label_for_single_image\nimport json\n\n\n# Custom Dataset Class\nclass CustomImageDataset(Dataset):\n def __init__(\n self, image_dir, transform=None, model_name=\"google/vit-base-patch16-224\"\n ):\n self.image_dir = image_dir\n self.transform = transform\n self.memory_bank = load_memory_bank() # Load existing labels from memory bank\n self.image_paths = [\n os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if img.endswith((\".png\", \".jpg\", \".jpeg\"))\n ]\n self.model_name = model_name\n self.label_to_index = {} # Dynamically map labels to indices\n self.current_index = 0 # Keep track of the next index to assign\n\n def __len__(self):\n return len(self.image_paths)\n\n def __getitem__(self, idx):\n image_path = self.image_paths[idx]\n image = Image.open(image_path).convert(\"RGB\")\n\n if image_path not in self.memory_bank: # Check if label already exists\n # Generate label for the image\n _, predicted_label = generate_label_for_single_image(image_path)\n self.memory_bank[image_path] = predicted_label # Add to memory bank\n else:\n predicted_label = self.memory_bank[image_path] # Use existing label\n # if self.transform:\n # image = self.transform(image)\n\n # Dynamically assign index to new labels\n if predicted_label not in self.label_to_index:\n self.label_to_index[predicted_label] = self.current_index\n self.current_index += 1\n label_index = self.label_to_index[predicted_label]\n\n if self.transform:\n image = self.transform(image)\n return image, torch.tensor(label_index, dtype=torch.long)\n\n\ndef load_memory_bank(filename=\"label_memory_bank.json\"):\n if os.path.exists(filename):\n with open(filename, \"r\") as file:\n return json.load(file)\n else:\n return {}","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:dataset.CustomImageDataset","uri":"program://SuperAlignment/class/dataset.CustomImageDataset#L10-L50","kind":"class","name":"CustomImageDataset","path":"dataset.py","language":"python","start_line":10,"end_line":50,"context_start_line":1,"context_end_line":58,"code":"import torch\nfrom torch.utils.data import Dataset\nimport os\nfrom PIL import Image\nfrom generate_labels import generate_label_for_single_image\nimport json\n\n\n# Custom Dataset Class\nclass CustomImageDataset(Dataset):\n def __init__(\n self, image_dir, transform=None, model_name=\"google/vit-base-patch16-224\"\n ):\n self.image_dir = image_dir\n self.transform = transform\n self.memory_bank = load_memory_bank() # Load existing labels from memory bank\n self.image_paths = [\n os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if img.endswith((\".png\", \".jpg\", \".jpeg\"))\n ]\n self.model_name = model_name\n self.label_to_index = {} # Dynamically map labels to indices\n self.current_index = 0 # Keep track of the next index to assign\n\n def __len__(self):\n return len(self.image_paths)\n\n def __getitem__(self, idx):\n image_path = self.image_paths[idx]\n image = Image.open(image_path).convert(\"RGB\")\n\n if image_path not in self.memory_bank: # Check if label already exists\n # Generate label for the image\n _, predicted_label = generate_label_for_single_image(image_path)\n self.memory_bank[image_path] = predicted_label # Add to memory bank\n else:\n predicted_label = self.memory_bank[image_path] # Use existing label\n # if self.transform:\n # image = self.transform(image)\n\n # Dynamically assign index to new labels\n if predicted_label not in self.label_to_index:\n self.label_to_index[predicted_label] = self.current_index\n self.current_index += 1\n label_index = self.label_to_index[predicted_label]\n\n if self.transform:\n image = self.transform(image)\n return image, torch.tensor(label_index, dtype=torch.long)\n\n\ndef load_memory_bank(filename=\"label_memory_bank.json\"):\n if os.path.exists(filename):\n with open(filename, \"r\") as file:\n return json.load(file)\n else:\n return {}","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:dataset.load_memory_bank","uri":"program://SuperAlignment/function/dataset.load_memory_bank#L53-L58","kind":"function","name":"load_memory_bank","path":"dataset.py","language":"python","start_line":53,"end_line":58,"context_start_line":33,"context_end_line":58,"code":" if image_path not in self.memory_bank: # Check if label already exists\n # Generate label for the image\n _, predicted_label = generate_label_for_single_image(image_path)\n self.memory_bank[image_path] = predicted_label # Add to memory bank\n else:\n predicted_label = self.memory_bank[image_path] # Use existing label\n # if self.transform:\n # image = self.transform(image)\n\n # Dynamically assign index to new labels\n if predicted_label not in self.label_to_index:\n self.label_to_index[predicted_label] = self.current_index\n self.current_index += 1\n label_index = self.label_to_index[predicted_label]\n\n if self.transform:\n image = self.transform(image)\n return image, torch.tensor(label_index, dtype=torch.long)\n\n\ndef load_memory_bank(filename=\"label_memory_bank.json\"):\n if os.path.exists(filename):\n with open(filename, \"r\") as file:\n return json.load(file)\n else:\n return {}","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:dataset.__init__","uri":"program://SuperAlignment/function/dataset.__init__#L11-L24","kind":"function","name":"__init__","path":"dataset.py","language":"python","start_line":11,"end_line":24,"context_start_line":1,"context_end_line":44,"code":"import torch\nfrom torch.utils.data import Dataset\nimport os\nfrom PIL import Image\nfrom generate_labels import generate_label_for_single_image\nimport json\n\n\n# Custom Dataset Class\nclass CustomImageDataset(Dataset):\n def __init__(\n self, image_dir, transform=None, model_name=\"google/vit-base-patch16-224\"\n ):\n self.image_dir = image_dir\n self.transform = transform\n self.memory_bank = load_memory_bank() # Load existing labels from memory bank\n self.image_paths = [\n os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if img.endswith((\".png\", \".jpg\", \".jpeg\"))\n ]\n self.model_name = model_name\n self.label_to_index = {} # Dynamically map labels to indices\n self.current_index = 0 # Keep track of the next index to assign\n\n def __len__(self):\n return len(self.image_paths)\n\n def __getitem__(self, idx):\n image_path = self.image_paths[idx]\n image = Image.open(image_path).convert(\"RGB\")\n\n if image_path not in self.memory_bank: # Check if label already exists\n # Generate label for the image\n _, predicted_label = generate_label_for_single_image(image_path)\n self.memory_bank[image_path] = predicted_label # Add to memory bank\n else:\n predicted_label = self.memory_bank[image_path] # Use existing label\n # if self.transform:\n # image = self.transform(image)\n\n # Dynamically assign index to new labels\n if predicted_label not in self.label_to_index:\n self.label_to_index[predicted_label] = self.current_index","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:dataset.__len__","uri":"program://SuperAlignment/function/dataset.__len__#L26-L27","kind":"function","name":"__len__","path":"dataset.py","language":"python","start_line":26,"end_line":27,"context_start_line":6,"context_end_line":47,"code":"import json\n\n\n# Custom Dataset Class\nclass CustomImageDataset(Dataset):\n def __init__(\n self, image_dir, transform=None, model_name=\"google/vit-base-patch16-224\"\n ):\n self.image_dir = image_dir\n self.transform = transform\n self.memory_bank = load_memory_bank() # Load existing labels from memory bank\n self.image_paths = [\n os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if img.endswith((\".png\", \".jpg\", \".jpeg\"))\n ]\n self.model_name = model_name\n self.label_to_index = {} # Dynamically map labels to indices\n self.current_index = 0 # Keep track of the next index to assign\n\n def __len__(self):\n return len(self.image_paths)\n\n def __getitem__(self, idx):\n image_path = self.image_paths[idx]\n image = Image.open(image_path).convert(\"RGB\")\n\n if image_path not in self.memory_bank: # Check if label already exists\n # Generate label for the image\n _, predicted_label = generate_label_for_single_image(image_path)\n self.memory_bank[image_path] = predicted_label # Add to memory bank\n else:\n predicted_label = self.memory_bank[image_path] # Use existing label\n # if self.transform:\n # image = self.transform(image)\n\n # Dynamically assign index to new labels\n if predicted_label not in self.label_to_index:\n self.label_to_index[predicted_label] = self.current_index\n self.current_index += 1\n label_index = self.label_to_index[predicted_label]\n","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:dataset.__getitem__","uri":"program://SuperAlignment/function/dataset.__getitem__#L29-L50","kind":"function","name":"__getitem__","path":"dataset.py","language":"python","start_line":29,"end_line":50,"context_start_line":9,"context_end_line":58,"code":"# Custom Dataset Class\nclass CustomImageDataset(Dataset):\n def __init__(\n self, image_dir, transform=None, model_name=\"google/vit-base-patch16-224\"\n ):\n self.image_dir = image_dir\n self.transform = transform\n self.memory_bank = load_memory_bank() # Load existing labels from memory bank\n self.image_paths = [\n os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if img.endswith((\".png\", \".jpg\", \".jpeg\"))\n ]\n self.model_name = model_name\n self.label_to_index = {} # Dynamically map labels to indices\n self.current_index = 0 # Keep track of the next index to assign\n\n def __len__(self):\n return len(self.image_paths)\n\n def __getitem__(self, idx):\n image_path = self.image_paths[idx]\n image = Image.open(image_path).convert(\"RGB\")\n\n if image_path not in self.memory_bank: # Check if label already exists\n # Generate label for the image\n _, predicted_label = generate_label_for_single_image(image_path)\n self.memory_bank[image_path] = predicted_label # Add to memory bank\n else:\n predicted_label = self.memory_bank[image_path] # Use existing label\n # if self.transform:\n # image = self.transform(image)\n\n # Dynamically assign index to new labels\n if predicted_label not in self.label_to_index:\n self.label_to_index[predicted_label] = self.current_index\n self.current_index += 1\n label_index = self.label_to_index[predicted_label]\n\n if self.transform:\n image = self.transform(image)\n return image, torch.tensor(label_index, dtype=torch.long)\n\n\ndef load_memory_bank(filename=\"label_memory_bank.json\"):\n if os.path.exists(filename):\n with open(filename, \"r\") as file:\n return json.load(file)\n else:\n return {}","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train","uri":"program://SuperAlignment/module/train#L1-L156","kind":"module","name":"train","path":"train.py","language":"python","start_line":1,"end_line":156,"context_start_line":1,"context_end_line":156,"code":"import torch\nimport torch.optim as optim\nfrom torchvision import transforms\nfrom torchvision.models import resnet18, resnet50\nfrom torch.utils.data import DataLoader\nimport numpy as np\nimport os\nimport json\nfrom utils import plot_and_save_metrics\nfrom dataset import CustomImageDataset\nfrom AdaptiveConfidenceDistillation import AdaptiveConfidenceDistillationLoss\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True) # Using pretrained=False for simplicity\nstudent_model = resnet50(pretrained=True)\n\n# Define a simple synthetic dataset resembling CIFAR-100\n# CIFAR-100 has 32x32 images with 3 channels\n# We'll create random data with similar dimensions\nnum_samples = 1000\ninput_channels = 3\ninput_height = 32\ninput_width = 32\nnum_classes = 100\n\n# Define transformations\ntransform = transforms.Compose(\n [\n transforms.Resize((32, 32)),\n transforms.ToTensor(),\n ]\n)\n\n# Create an instance of the CustomImageDataset\nimage_dir = \"./images\" # Specify the correct path to your images\ndataset = CustomImageDataset(image_dir=image_dir, transform=transform)\n\n\ndef save_memory_bank(memory_bank, filename=\"label_memory_bank.json\"):\n with open(filename, \"w\") as file:\n json.dump(memory_bank, file)\n\n\n# After training is complete or when you're done using the dataset\nsave_memory_bank(dataset.memory_bank) # Save the updated memory bank to a file\n# Split dataset into train and validation sets\ntrain_size = int(0.8 * len(dataset))\nval_size = len(dataset) - train_size\ntrain_dataset, val_dataset = torch.utils.data.random_split(\n dataset, [train_size, val_size]\n)\n\n# Create DataLoaders\ntrain_dataloader = DataLoader(train_dataset, batch_size=32, shuffle=True)\nval_dataloader = DataLoader(val_dataset, batch_size=32, shuffle=False)\n\n# Define optimizer and device\noptimizer = optim.Adam(student_model.parameters(), lr=0.001)\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\nprint(f\"Using device: {device}\")\n\n\n# Step 1: Load Labels and Image Paths\ndef load_image_paths_labels(json_path):\n with open(json_path, \"r\") as f:\n data = json.load(f)\n image_paths = [os.path.join(\"./images\", item[\"image_path\"]) for item in data]\n labels = [item[\"caption\"] for item in data]\n return image_paths, labels\n\n\n# Example Training Loop Skeleton with Validation and Model Checkpointing\ndef train_model_with_validation(\n student_model,\n teacher_model,\n train_dataloader,\n val_dataloader,\n optimizer,\n device,\n num_epochs=10,\n checkpoint_path=\"checkpoint.pth\",\n):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n # Initialize metrics dictionary\n metrics = {\n \"train_loss\": [],\n \"val_loss\": [],\n } # Add 'train_accuracy': [], 'val_accuracy': [] if tracking accuracy\n loss_fn = AdaptiveConfidenceDistillationLoss(temperature=2.0)\n best_val_loss = float(\"inf\")\n step = 0\n for epoch in range(num_epochs):\n # Training loop\n for inputs, labels in train_dataloader:\n if step % 200 == 0:\n # Validation loop\n val_loss = 0.0\n with torch.no_grad():\n for inputs, labels in val_dataloader:\n inputs, labels = inputs.to(device), labels.to(device)\n teacher_outputs = teacher_model(inputs)\n student_outputs = student_model(inputs)\n loss = loss_fn(student_outputs, teacher_outputs, labels)\n val_loss += loss.item() * inputs.size(\n 0\n ) # Inside the training loop\n metrics[\"train_loss\"].append(\n loss.item()\n ) # Assuming loss is your loss variable\n\n # Inside the validation loop\n metrics[\"val_loss\"].append(\n val_loss\n ) # Assuming val_loss is your accumulated validation loss\n step += 1\n print(f\"val_loss: {val_loss}\")\n # After training, plot and save metrics\n plot_and_save_metrics(metrics)\n inputs, labels = inputs.to(device), labels.to(device)\n optimizer.zero_grad()\n\n with torch.no_grad():\n teacher_outputs = teacher_model(inputs)\n\n student_outputs = student_model(inputs)\n\n loss = loss_fn(student_outputs, teacher_outputs, labels)\n loss.backward()\n optimizer.step()\n\n val_loss /= len(val_dataloader.dataset)\n print(f\"Epoch [{epoch+1}/{num_epochs}], Validation Loss: {val_loss:.4f}\")\n\n # Save the model checkpoint if validation loss improves\n if val_loss < best_val_loss:\n best_val_loss = val_loss\n torch.save(student_model.state_dict(), checkpoint_path)\n # After training, plot and save metrics\n plot_and_save_metrics(metrics)\n print(f\"Saving model checkpoint at {checkpoint_path}\")\n\n\n# Example usage\n# Assuming `train_dataloader`, `val_dataloader`, `optimizer`, and `device` are defined\ntrain_model_with_validation(\n student_model,\n teacher_model,\n train_dataloader,\n val_dataloader,\n optimizer,\n device,\n num_epochs=10,\n checkpoint_path=\"checkpoint.pth\",\n)","source_hash":"aec049768a636ee5c1da4badd75c577ef9b0a0665268f1006af96fee0e488a23","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train.save_memory_bank","uri":"program://SuperAlignment/function/train.save_memory_bank#L39-L41","kind":"function","name":"save_memory_bank","path":"train.py","language":"python","start_line":39,"end_line":41,"context_start_line":19,"context_end_line":61,"code":"# We'll create random data with similar dimensions\nnum_samples = 1000\ninput_channels = 3\ninput_height = 32\ninput_width = 32\nnum_classes = 100\n\n# Define transformations\ntransform = transforms.Compose(\n [\n transforms.Resize((32, 32)),\n transforms.ToTensor(),\n ]\n)\n\n# Create an instance of the CustomImageDataset\nimage_dir = \"./images\" # Specify the correct path to your images\ndataset = CustomImageDataset(image_dir=image_dir, transform=transform)\n\n\ndef save_memory_bank(memory_bank, filename=\"label_memory_bank.json\"):\n with open(filename, \"w\") as file:\n json.dump(memory_bank, file)\n\n\n# After training is complete or when you're done using the dataset\nsave_memory_bank(dataset.memory_bank) # Save the updated memory bank to a file\n# Split dataset into train and validation sets\ntrain_size = int(0.8 * len(dataset))\nval_size = len(dataset) - train_size\ntrain_dataset, val_dataset = torch.utils.data.random_split(\n dataset, [train_size, val_size]\n)\n\n# Create DataLoaders\ntrain_dataloader = DataLoader(train_dataset, batch_size=32, shuffle=True)\nval_dataloader = DataLoader(val_dataset, batch_size=32, shuffle=False)\n\n# Define optimizer and device\noptimizer = optim.Adam(student_model.parameters(), lr=0.001)\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\nprint(f\"Using device: {device}\")\n","source_hash":"aec049768a636ee5c1da4badd75c577ef9b0a0665268f1006af96fee0e488a23","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train.load_image_paths_labels","uri":"program://SuperAlignment/function/train.load_image_paths_labels#L64-L69","kind":"function","name":"load_image_paths_labels","path":"train.py","language":"python","start_line":64,"end_line":69,"context_start_line":44,"context_end_line":89,"code":"# After training is complete or when you're done using the dataset\nsave_memory_bank(dataset.memory_bank) # Save the updated memory bank to a file\n# Split dataset into train and validation sets\ntrain_size = int(0.8 * len(dataset))\nval_size = len(dataset) - train_size\ntrain_dataset, val_dataset = torch.utils.data.random_split(\n dataset, [train_size, val_size]\n)\n\n# Create DataLoaders\ntrain_dataloader = DataLoader(train_dataset, batch_size=32, shuffle=True)\nval_dataloader = DataLoader(val_dataset, batch_size=32, shuffle=False)\n\n# Define optimizer and device\noptimizer = optim.Adam(student_model.parameters(), lr=0.001)\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\nprint(f\"Using device: {device}\")\n\n\n# Step 1: Load Labels and Image Paths\ndef load_image_paths_labels(json_path):\n with open(json_path, \"r\") as f:\n data = json.load(f)\n image_paths = [os.path.join(\"./images\", item[\"image_path\"]) for item in data]\n labels = [item[\"caption\"] for item in data]\n return image_paths, labels\n\n\n# Example Training Loop Skeleton with Validation and Model Checkpointing\ndef train_model_with_validation(\n student_model,\n teacher_model,\n train_dataloader,\n val_dataloader,\n optimizer,\n device,\n num_epochs=10,\n checkpoint_path=\"checkpoint.pth\",\n):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n # Initialize metrics dictionary\n metrics = {\n \"train_loss\": [],\n \"val_loss\": [],","source_hash":"aec049768a636ee5c1da4badd75c577ef9b0a0665268f1006af96fee0e488a23","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:train.train_model_with_validation","uri":"program://SuperAlignment/function/train.train_model_with_validation#L73-L142","kind":"function","name":"train_model_with_validation","path":"train.py","language":"python","start_line":73,"end_line":142,"context_start_line":53,"context_end_line":156,"code":"# Create DataLoaders\ntrain_dataloader = DataLoader(train_dataset, batch_size=32, shuffle=True)\nval_dataloader = DataLoader(val_dataset, batch_size=32, shuffle=False)\n\n# Define optimizer and device\noptimizer = optim.Adam(student_model.parameters(), lr=0.001)\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\nprint(f\"Using device: {device}\")\n\n\n# Step 1: Load Labels and Image Paths\ndef load_image_paths_labels(json_path):\n with open(json_path, \"r\") as f:\n data = json.load(f)\n image_paths = [os.path.join(\"./images\", item[\"image_path\"]) for item in data]\n labels = [item[\"caption\"] for item in data]\n return image_paths, labels\n\n\n# Example Training Loop Skeleton with Validation and Model Checkpointing\ndef train_model_with_validation(\n student_model,\n teacher_model,\n train_dataloader,\n val_dataloader,\n optimizer,\n device,\n num_epochs=10,\n checkpoint_path=\"checkpoint.pth\",\n):\n student_model.train()\n teacher_model.eval() # Teacher model should be in eval mode\n\n # Initialize metrics dictionary\n metrics = {\n \"train_loss\": [],\n \"val_loss\": [],\n } # Add 'train_accuracy': [], 'val_accuracy': [] if tracking accuracy\n loss_fn = AdaptiveConfidenceDistillationLoss(temperature=2.0)\n best_val_loss = float(\"inf\")\n step = 0\n for epoch in range(num_epochs):\n # Training loop\n for inputs, labels in train_dataloader:\n if step % 200 == 0:\n # Validation loop\n val_loss = 0.0\n with torch.no_grad():\n for inputs, labels in val_dataloader:\n inputs, labels = inputs.to(device), labels.to(device)\n teacher_outputs = teacher_model(inputs)\n student_outputs = student_model(inputs)\n loss = loss_fn(student_outputs, teacher_outputs, labels)\n val_loss += loss.item() * inputs.size(\n 0\n ) # Inside the training loop\n metrics[\"train_loss\"].append(\n loss.item()\n ) # Assuming loss is your loss variable\n\n # Inside the validation loop\n metrics[\"val_loss\"].append(\n val_loss\n ) # Assuming val_loss is your accumulated validation loss\n step += 1\n print(f\"val_loss: {val_loss}\")\n # After training, plot and save metrics\n plot_and_save_metrics(metrics)\n inputs, labels = inputs.to(device), labels.to(device)\n optimizer.zero_grad()\n\n with torch.no_grad():\n teacher_outputs = teacher_model(inputs)\n\n student_outputs = student_model(inputs)\n\n loss = loss_fn(student_outputs, teacher_outputs, labels)\n loss.backward()\n optimizer.step()\n\n val_loss /= len(val_dataloader.dataset)\n print(f\"Epoch [{epoch+1}/{num_epochs}], Validation Loss: {val_loss:.4f}\")\n\n # Save the model checkpoint if validation loss improves\n if val_loss < best_val_loss:\n best_val_loss = val_loss\n torch.save(student_model.state_dict(), checkpoint_path)\n # After training, plot and save metrics\n plot_and_save_metrics(metrics)\n print(f\"Saving model checkpoint at {checkpoint_path}\")\n\n\n# Example usage\n# Assuming `train_dataloader`, `val_dataloader`, `optimizer`, and `device` are defined\ntrain_model_with_validation(\n student_model,\n teacher_model,\n train_dataloader,\n val_dataloader,\n optimizer,\n device,\n num_epochs=10,\n checkpoint_path=\"checkpoint.pth\",\n)","source_hash":"aec049768a636ee5c1da4badd75c577ef9b0a0665268f1006af96fee0e488a23","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:utils","uri":"program://SuperAlignment/module/utils#L1-L34","kind":"module","name":"utils","path":"utils.py","language":"python","start_line":1,"end_line":34,"context_start_line":1,"context_end_line":34,"code":"import matplotlib.pyplot as plt\nimport json\n\n\ndef plot_and_save_metrics(\n metrics, filename=\"training_metrics.png\", metrics_path=\"training_metrics.json\"\n):\n epochs = range(1, len(metrics[\"train_loss\"]) + 1)\n\n plt.figure(figsize=(10, 5))\n plt.subplot(1, 2, 1)\n plt.plot(epochs, metrics[\"train_loss\"], label=\"Training Loss\")\n plt.plot(epochs, metrics[\"val_loss\"], label=\"Validation Loss\")\n plt.title(\"Loss Metrics\")\n plt.xlabel(\"Epoch\")\n plt.ylabel(\"Loss\")\n plt.legend()\n\n # If you track accuracy, you can plot it here similarly.\n # plt.subplot(1, 2, 2)\n # plt.plot(epochs, metrics['train_accuracy'], label='Training Accuracy')\n # plt.plot(epochs, metrics['val_accuracy'], label='Validation Accuracy')\n # plt.title('Accuracy Metrics')\n # plt.xlabel('Epoch')\n # plt.ylabel('Accuracy')\n # plt.legend()\n\n plt.tight_layout()\n plt.savefig(filename)\n plt.close()\n\n # Save metrics to a JSON file for further analysis\n with open(metrics_path, \"w\") as f:\n json.dump(metrics, f, indent=4)","source_hash":"7804c1866cae172702f44d671accd34ee657dc7840952f9d6700cc17f267e8bd","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:utils.plot_and_save_metrics","uri":"program://SuperAlignment/function/utils.plot_and_save_metrics#L5-L34","kind":"function","name":"plot_and_save_metrics","path":"utils.py","language":"python","start_line":5,"end_line":34,"context_start_line":1,"context_end_line":34,"code":"import matplotlib.pyplot as plt\nimport json\n\n\ndef plot_and_save_metrics(\n metrics, filename=\"training_metrics.png\", metrics_path=\"training_metrics.json\"\n):\n epochs = range(1, len(metrics[\"train_loss\"]) + 1)\n\n plt.figure(figsize=(10, 5))\n plt.subplot(1, 2, 1)\n plt.plot(epochs, metrics[\"train_loss\"], label=\"Training Loss\")\n plt.plot(epochs, metrics[\"val_loss\"], label=\"Validation Loss\")\n plt.title(\"Loss Metrics\")\n plt.xlabel(\"Epoch\")\n plt.ylabel(\"Loss\")\n plt.legend()\n\n # If you track accuracy, you can plot it here similarly.\n # plt.subplot(1, 2, 2)\n # plt.plot(epochs, metrics['train_accuracy'], label='Training Accuracy')\n # plt.plot(epochs, metrics['val_accuracy'], label='Validation Accuracy')\n # plt.title('Accuracy Metrics')\n # plt.xlabel('Epoch')\n # plt.ylabel('Accuracy')\n # plt.legend()\n\n plt.tight_layout()\n plt.savefig(filename)\n plt.close()\n\n # Save metrics to a JSON file for further analysis\n with open(metrics_path, \"w\") as f:\n json.dump(metrics, f, indent=4)","source_hash":"7804c1866cae172702f44d671accd34ee657dc7840952f9d6700cc17f267e8bd","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:generate_labels","uri":"program://SuperAlignment/module/generate_labels#L1-L78","kind":"module","name":"generate_labels","path":"generate_labels.py","language":"python","start_line":1,"end_line":78,"context_start_line":1,"context_end_line":78,"code":"import os\nfrom transformers import ViTImageProcessor, ViTForImageClassification\nfrom PIL import Image\nimport json\nimport torch\n\n\ndef generate_label_for_single_image(image_path):\n # Initialize the model and processor\n model = ViTForImageClassification.from_pretrained(\"google/vit-base-patch16-224\")\n feature_extractor = ViTImageProcessor.from_pretrained(\"google/vit-base-patch16-224\")\n\n # Load and process the image\n image = Image.open(image_path).convert(\"RGB\")\n inputs = feature_extractor(images=image, return_tensors=\"pt\")\n\n # Predict the class of the image\n with torch.no_grad():\n outputs = model(**inputs)\n\n logits = outputs.logits\n predicted_class_idx = logits.argmax(-1).item()\n\n # Optionally, print the predicted class (if id2label is available)\n predicted_label = (\n model.config.id2label[predicted_class_idx]\n if model.config.id2label\n else str(predicted_class_idx)\n )\n\n print(\"Predicted class index:\", predicted_class_idx)\n print(\"Predicted label:\", predicted_label)\n\n return predicted_class_idx, predicted_label\n\n\ndef generate_labels(image_dir, model_name, output_file, max_images=100):\n # Load the CLIP model and processor\n processor = ViTImageProcessor.from_pretrained(\"google/vit-base-patch16-224\")\n model = ViTForImageClassification.from_pretrained(\"google/vit-base-patch16-224\")\n\n # Fetch image paths\n images = os.listdir(image_dir)\n images = images[:max_images] # Limit to max_images\n\n labels = []\n\n for image_name in images:\n image_path = os.path.join(image_dir, image_name)\n\n # Load image\n image = Image.open(image_path).convert(\"RGB\")\n\n # Prepare the image for the model\n inputs = processor(images=image, return_tensors=\"pt\")\n\n # Generate image features (embeddings)\n with torch.no_grad():\n outputs = model(**inputs)\n\n # Here you would typically use the embeddings to find the closest text descriptions, but since\n # OpenCLIP doesn't provide direct label generation, we'll just save the embeddings for now.\n # This step is a placeholder for any specific label generation or nearest neighbor search you might implement.\n # For demonstration, let's assume a dummy label based on the max value index in the embeddings.\n logits = outputs.logits\n # model predicts one of the 1000 ImageNet classes\n predicted_class_idx = logits.argmax(-1).item()\n\n labels.append({\"image_path\": image_name, \"label\": predicted_class_idx})\n print(\"Predicted class: \", model.config.id2label[predicted_class_idx])\n # Save the labels to a JSON file\n with open(output_file, \"w\") as f:\n json.dump(labels, f, indent=4)\n\n\nimage_dir = \"./images\"\nmodel_name = \"google/vit-base-patch16-224\"\noutput_file = \"imagelabels.json\"","source_hash":"20a6fcb44164bd9974723b003b234785c2e029411323ddda3cb2fab53143ef64","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:generate_labels.generate_label_for_single_image","uri":"program://SuperAlignment/function/generate_labels.generate_label_for_single_image#L8-L34","kind":"function","name":"generate_label_for_single_image","path":"generate_labels.py","language":"python","start_line":8,"end_line":34,"context_start_line":1,"context_end_line":54,"code":"import os\nfrom transformers import ViTImageProcessor, ViTForImageClassification\nfrom PIL import Image\nimport json\nimport torch\n\n\ndef generate_label_for_single_image(image_path):\n # Initialize the model and processor\n model = ViTForImageClassification.from_pretrained(\"google/vit-base-patch16-224\")\n feature_extractor = ViTImageProcessor.from_pretrained(\"google/vit-base-patch16-224\")\n\n # Load and process the image\n image = Image.open(image_path).convert(\"RGB\")\n inputs = feature_extractor(images=image, return_tensors=\"pt\")\n\n # Predict the class of the image\n with torch.no_grad():\n outputs = model(**inputs)\n\n logits = outputs.logits\n predicted_class_idx = logits.argmax(-1).item()\n\n # Optionally, print the predicted class (if id2label is available)\n predicted_label = (\n model.config.id2label[predicted_class_idx]\n if model.config.id2label\n else str(predicted_class_idx)\n )\n\n print(\"Predicted class index:\", predicted_class_idx)\n print(\"Predicted label:\", predicted_label)\n\n return predicted_class_idx, predicted_label\n\n\ndef generate_labels(image_dir, model_name, output_file, max_images=100):\n # Load the CLIP model and processor\n processor = ViTImageProcessor.from_pretrained(\"google/vit-base-patch16-224\")\n model = ViTForImageClassification.from_pretrained(\"google/vit-base-patch16-224\")\n\n # Fetch image paths\n images = os.listdir(image_dir)\n images = images[:max_images] # Limit to max_images\n\n labels = []\n\n for image_name in images:\n image_path = os.path.join(image_dir, image_name)\n\n # Load image\n image = Image.open(image_path).convert(\"RGB\")\n\n # Prepare the image for the model","source_hash":"20a6fcb44164bd9974723b003b234785c2e029411323ddda3cb2fab53143ef64","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"py:generate_labels.generate_labels","uri":"program://SuperAlignment/function/generate_labels.generate_labels#L37-L73","kind":"function","name":"generate_labels","path":"generate_labels.py","language":"python","start_line":37,"end_line":73,"context_start_line":17,"context_end_line":78,"code":" # Predict the class of the image\n with torch.no_grad():\n outputs = model(**inputs)\n\n logits = outputs.logits\n predicted_class_idx = logits.argmax(-1).item()\n\n # Optionally, print the predicted class (if id2label is available)\n predicted_label = (\n model.config.id2label[predicted_class_idx]\n if model.config.id2label\n else str(predicted_class_idx)\n )\n\n print(\"Predicted class index:\", predicted_class_idx)\n print(\"Predicted label:\", predicted_label)\n\n return predicted_class_idx, predicted_label\n\n\ndef generate_labels(image_dir, model_name, output_file, max_images=100):\n # Load the CLIP model and processor\n processor = ViTImageProcessor.from_pretrained(\"google/vit-base-patch16-224\")\n model = ViTForImageClassification.from_pretrained(\"google/vit-base-patch16-224\")\n\n # Fetch image paths\n images = os.listdir(image_dir)\n images = images[:max_images] # Limit to max_images\n\n labels = []\n\n for image_name in images:\n image_path = os.path.join(image_dir, image_name)\n\n # Load image\n image = Image.open(image_path).convert(\"RGB\")\n\n # Prepare the image for the model\n inputs = processor(images=image, return_tensors=\"pt\")\n\n # Generate image features (embeddings)\n with torch.no_grad():\n outputs = model(**inputs)\n\n # Here you would typically use the embeddings to find the closest text descriptions, but since\n # OpenCLIP doesn't provide direct label generation, we'll just save the embeddings for now.\n # This step is a placeholder for any specific label generation or nearest neighbor search you might implement.\n # For demonstration, let's assume a dummy label based on the max value index in the embeddings.\n logits = outputs.logits\n # model predicts one of the 1000 ImageNet classes\n predicted_class_idx = logits.argmax(-1).item()\n\n labels.append({\"image_path\": image_name, \"label\": predicted_class_idx})\n print(\"Predicted class: \", model.config.id2label[predicted_class_idx])\n # Save the labels to a JSON file\n with open(output_file, \"w\") as f:\n json.dump(labels, f, indent=4)\n\n\nimage_dir = \"./images\"\nmodel_name = \"google/vit-base-patch16-224\"\noutput_file = \"imagelabels.json\"","source_hash":"20a6fcb44164bd9974723b003b234785c2e029411323ddda3cb2fab53143ef64","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:train_example.py","uri":"program://SuperAlignment/file/train_example.py","kind":"file","name":"train_example.py","path":"train_example.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nfrom torchvision.models import resnet18, resnet50 # Example models\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True)\nstudent_model = resnet50(pretrained=True)\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)","source_hash":"1d853ce261864b7b90ca11a013e5ff6feff4aa1e87a9d5d20878083564c9aba6","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:AdaptiveConfidenceDistillation.py","uri":"program://SuperAlignment/file/AdaptiveConfidenceDistillation.py","kind":"file","name":"AdaptiveConfidenceDistillation.py","path":"AdaptiveConfidenceDistillation.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import torch.nn as nn\nimport torch.nn.functional as F\n\n\n# Custom Adaptive Confidence Distillation Loss\nclass AdaptiveConfidenceDistillationLoss(nn.Module):\n def __init__(self, temperature=1.0):\n super().__init__()\n self.temperature = temperature\n self.ce_loss = nn.CrossEntropyLoss()\n\n def forward(self, student_logits, teacher_logits, targets):\n # Soften probabilities\n soft_teacher_probs = F.softmax(teacher_logits / self.temperature, dim=1)\n soft_student_probs = F.log_softmax(student_logits / self.temperature, dim=1)\n\n # Calculate the KL Divergence for the soft targets\n distillation_loss = F.kl_div(\n soft_student_probs, soft_teacher_probs, reduction=\"batchmean\"\n )\n","source_hash":"78f2354cb5ad0726bbf11cff4469be7a070132d3339921881ae5df2398b71f16","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:test_train_example.py","uri":"program://SuperAlignment/file/test_train_example.py","kind":"file","name":"test_train_example.py","path":"test_train_example.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torchvision.models import resnet18, resnet50\nfrom torch.utils.data import DataLoader, TensorDataset\nimport numpy as np\nfrom train_example import train_model\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=False) # Using pretrained=False for simplicity\nstudent_model = resnet50(pretrained=False)\n\n# Define a simple synthetic dataset resembling CIFAR-100\n# CIFAR-100 has 32x32 images with 3 channels\n# We'll create random data with similar dimensions\nnum_samples = 1000\ninput_channels = 3\ninput_height = 32\ninput_width = 32\nnum_classes = 100\n","source_hash":"6e808ad91b17c4e12df91f6ed188bfe810f9d6cb3ef55c61df0055a57b0abc0b","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:dataset.py","uri":"program://SuperAlignment/file/dataset.py","kind":"file","name":"dataset.py","path":"dataset.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import torch\nfrom torch.utils.data import Dataset\nimport os\nfrom PIL import Image\nfrom generate_labels import generate_label_for_single_image\nimport json\n\n\n# Custom Dataset Class\nclass CustomImageDataset(Dataset):\n def __init__(\n self, image_dir, transform=None, model_name=\"google/vit-base-patch16-224\"\n ):\n self.image_dir = image_dir\n self.transform = transform\n self.memory_bank = load_memory_bank() # Load existing labels from memory bank\n self.image_paths = [\n os.path.join(image_dir, img)\n for img in os.listdir(image_dir)\n if img.endswith((\".png\", \".jpg\", \".jpeg\"))\n ]","source_hash":"981d68dd2e2858f57d74c1cd302ee09901a2fa38c665753fa9e6a5a0b08f1267","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:train.py","uri":"program://SuperAlignment/file/train.py","kind":"file","name":"train.py","path":"train.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import torch\nimport torch.optim as optim\nfrom torchvision import transforms\nfrom torchvision.models import resnet18, resnet50\nfrom torch.utils.data import DataLoader\nimport numpy as np\nimport os\nimport json\nfrom utils import plot_and_save_metrics\nfrom dataset import CustomImageDataset\nfrom AdaptiveConfidenceDistillation import AdaptiveConfidenceDistillationLoss\n\n# Define the teacher (weak model) and student (strong model)\nteacher_model = resnet18(pretrained=True) # Using pretrained=False for simplicity\nstudent_model = resnet50(pretrained=True)\n\n# Define a simple synthetic dataset resembling CIFAR-100\n# CIFAR-100 has 32x32 images with 3 channels\n# We'll create random data with similar dimensions\nnum_samples = 1000\ninput_channels = 3","source_hash":"aec049768a636ee5c1da4badd75c577ef9b0a0665268f1006af96fee0e488a23","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:utils.py","uri":"program://SuperAlignment/file/utils.py","kind":"file","name":"utils.py","path":"utils.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import matplotlib.pyplot as plt\nimport json\n\n\ndef plot_and_save_metrics(\n metrics, filename=\"training_metrics.png\", metrics_path=\"training_metrics.json\"\n):\n epochs = range(1, len(metrics[\"train_loss\"]) + 1)\n\n plt.figure(figsize=(10, 5))\n plt.subplot(1, 2, 1)\n plt.plot(epochs, metrics[\"train_loss\"], label=\"Training Loss\")\n plt.plot(epochs, metrics[\"val_loss\"], label=\"Validation Loss\")\n plt.title(\"Loss Metrics\")\n plt.xlabel(\"Epoch\")\n plt.ylabel(\"Loss\")\n plt.legend()\n\n # If you track accuracy, you can plot it here similarly.\n # plt.subplot(1, 2, 2)\n # plt.plot(epochs, metrics['train_accuracy'], label='Training Accuracy')","source_hash":"7804c1866cae172702f44d671accd34ee657dc7840952f9d6700cc17f267e8bd","truncated":false}
{"repo_id":"SuperAlignment","entity_id":"file:generate_labels.py","uri":"program://SuperAlignment/file/generate_labels.py","kind":"file","name":"generate_labels.py","path":"generate_labels.py","language":"python","start_line":1,"end_line":1,"context_start_line":1,"context_end_line":21,"code":"import os\nfrom transformers import ViTImageProcessor, ViTForImageClassification\nfrom PIL import Image\nimport json\nimport torch\n\n\ndef generate_label_for_single_image(image_path):\n # Initialize the model and processor\n model = ViTForImageClassification.from_pretrained(\"google/vit-base-patch16-224\")\n feature_extractor = ViTImageProcessor.from_pretrained(\"google/vit-base-patch16-224\")\n\n # Load and process the image\n image = Image.open(image_path).convert(\"RGB\")\n inputs = feature_extractor(images=image, return_tensors=\"pt\")\n\n # Predict the class of the image\n with torch.no_grad():\n outputs = model(**inputs)\n\n logits = outputs.logits","source_hash":"20a6fcb44164bd9974723b003b234785c2e029411323ddda3cb2fab53143ef64","truncated":false}