# server.py # Server Flask untuk prediksi multi-head captcha yang disesuaikan untuk Hugging Face Spaces. # Versi ini sudah dioptimalkan untuk memuat model yang terkuantisasi (INT8). from flask import Flask, request, jsonify from flask_cors import CORS import torch import torch.nn as nn import timm import albumentations as A from albumentations.pytorch import ToTensorV2 from PIL import Image, UnidentifiedImageError import numpy as np import cv2 import os import json import base64 from io import BytesIO import sys import logging torch.set_num_threads(1) # ============================================================================== # BAGIAN 1: PENGATURAN DASAR FLASK & LOGGING # ============================================================================== logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s', stream=sys.stdout) app = Flask(__name__) CORS(app) MODEL = None MAPPINGS = None DEVICE = None TRANSFORMS = None # ============================================================================== # BAGIAN 2: DEFINISI MODEL (TIDAK ADA PERUBAHAN) # ============================================================================== class TextHeadCTC(nn.Module): def __init__(self, input_dim, hidden_dim, ctc_vocab_size): super().__init__() self.rnn = nn.LSTM(input_dim, hidden_dim, num_layers=1, bidirectional=True, batch_first=True) self.fc = nn.Linear(hidden_dim * 2, ctc_vocab_size) def forward(self, x): rnn_out, _ = self.rnn(x) output_logits = self.fc(rnn_out) return nn.functional.log_softmax(output_logits, dim=2).permute(1, 0, 2) class MultiHeadModel(nn.Module): def __init__(self, backbone_name, ctc_vocab_size, num_object_classes, num_types): super().__init__() self.backbone = timm.create_model( backbone_name, pretrained=False, num_classes=0, drop_path_rate=0.1 ) backbone_features_dim = self.backbone.num_features rnn_hidden_dim, projected_embed_dim = 256, 256 self.type_head = nn.Linear(backbone_features_dim, num_types) self.object_head = nn.Linear(backbone_features_dim, num_object_classes) self.input_proj = nn.Conv2d(backbone_features_dim, projected_embed_dim, kernel_size=1) self.text_head_ctc = TextHeadCTC(projected_embed_dim, rnn_hidden_dim, ctc_vocab_size) self.pool = nn.AdaptiveAvgPool2d((1, 1)) def forward(self, x): features = self.backbone.forward_features(x) pooled_features = self.pool(features).flatten(1) type_logits = self.type_head(pooled_features) object_logits = self.object_head(pooled_features) proj_features = self.input_proj(features) bs, c_proj, h_feat, w_feat = proj_features.size() image_features_seq = proj_features.view(bs, c_proj, h_feat * w_feat).permute(0, 2, 1) text_log_probs = self.text_head_ctc(image_features_seq) return type_logits, object_logits, text_log_probs # ============================================================================== # BAGIAN 3: FUNGSI HELPER (TIDAK ADA PERUBAHAN) # ============================================================================== def get_transforms(img_height, img_width): interpolation_method = cv2.INTER_AREA return A.Compose([ A.Resize(height=img_height, width=img_width, interpolation=interpolation_method), A.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]), ToTensorV2() ]) def ctc_decoder_with_confidence(log_probs, idx_to_char_map, blank_idx): probs = torch.exp(log_probs) max_probs, pred_indices = torch.max(probs, dim=-1) max_probs = max_probs.squeeze(1).cpu().numpy() pred_indices = pred_indices.squeeze(1).cpu().numpy() decoded_sequence = [] confidence_values = [] last_idx = -1 for i, idx in enumerate(pred_indices): if idx == blank_idx or idx == last_idx: last_idx = blank_idx if idx == blank_idx else last_idx continue decoded_sequence.append(idx_to_char_map.get(str(idx), '?')) confidence_values.append(max_probs[i]) last_idx = idx final_text = "".join(decoded_sequence) avg_confidence = np.mean(confidence_values) if confidence_values else 0.0 return final_text, avg_confidence # ============================================================================== # BAGIAN 4: INISIALISASI SERVER (VERSI UNTUK MODEL KUANTISASI) # ============================================================================== def initialize_server(model_path, mappings_path): global MODEL, MAPPINGS, DEVICE, TRANSFORMS logging.info("Memulai inisialisasi server dengan model terkuantisasi...") DEVICE = torch.device("cpu") # Kuantisasi INT8 dioptimalkan untuk CPU logging.info(f"Menggunakan device: {DEVICE}") try: if not os.path.exists(mappings_path): raise FileNotFoundError(f"File mappings tidak ditemukan di: {mappings_path}") with open(mappings_path, 'r', encoding='utf-8') as f: MAPPINGS = json.load(f) logging.info("File mappings berhasil dimuat.") except Exception as e: logging.error(f"FATAL: Gagal memuat file mappings: {e}") sys.exit(1) TRANSFORMS = get_transforms(MAPPINGS['img_height'], MAPPINGS['img_width']) # Langkah 1: Buat instance model dengan arsitektur asli try: m = MAPPINGS model_to_quantize = MultiHeadModel( backbone_name=m['backbone'], ctc_vocab_size=len(m['ctc_char_to_idx']), num_object_classes=len(m['object_to_idx']), num_types=len(m['type_to_idx']) ) logging.info(f"Instance model '{MAPPINGS['backbone']}' berhasil dibuat.") except Exception as e: logging.error(f"FATAL: Gagal membuat instance model. Error: {e}") sys.exit(1) # Langkah 2: Siapkan model untuk menerima weights terkuantisasi MODEL = torch.quantization.quantize_dynamic( model_to_quantize, {nn.Linear, nn.LSTM}, dtype=torch.qint8 ) logging.info("Arsitektur model disiapkan untuk kuantisasi dinamis.") # Langkah 3: Muat state_dict dari file .pth yang sudah terkuantisasi try: if not os.path.exists(model_path): raise FileNotFoundError(f"File model tidak ditemukan di: {model_path}") # Langsung load state_dict karena kita sudah menyiapkan arsitekturnya MODEL.load_state_dict(torch.load(model_path, map_location=DEVICE)) MODEL.to(DEVICE) MODEL.eval() logging.info("Model weights terkuantisasi berhasil dimuat dan siap digunakan.") except Exception as e: logging.error(f"FATAL: Gagal memuat model weights terkuantisasi. Error: {e}", exc_info=True) sys.exit(1) logging.info("Inisialisasi server selesai. Siap menerima permintaan.") # ============================================================================== # BAGIAN 5: ENDPOINT FLASK (TIDAK ADA PERUBAHAN) # ============================================================================== @app.route('/', methods=['GET']) def home(): """Endpoint dasar untuk memeriksa apakah server berjalan.""" return "

Captcha Prediction Server is running (Quantized Model).

Gunakan endpoint /predict untuk melakukan prediksi.

", 200 @app.route('/predict', methods=['POST']) def predict_endpoint(): """Endpoint untuk menerima gambar base64 dan mengembalikan prediksi.""" # Otentikasi expected_api_key = os.environ.get('API_KEY_SECRET') if not expected_api_key: logging.error("FATAL: Secret 'API_KEY_SECRET' tidak diatur di server.") return jsonify({"error": "Konfigurasi server error."}), 500 auth_header = request.headers.get('Authorization') if not auth_header or auth_header != f"Bearer {expected_api_key}": logging.warning(f"Akses ditolak untuk IP: {request.remote_addr}. Alasan: Kunci API tidak valid.") return jsonify({"error": "Akses ditolak."}), 403 # Proses prediksi if not request.is_json: return jsonify({"error": "Request harus berupa JSON"}), 400 data = request.get_json() base64_string = data.get('image_base64') if not base64_string: return jsonify({"error": "Key 'image_base64' tidak ditemukan atau kosong"}), 400 try: if ',' in base64_string: _, encoded = base64_string.split(',', 1) else: encoded = base64_string image_data = base64.b64decode(encoded) img_pil = Image.open(BytesIO(image_data)).convert("RGB") except (base64.binascii.Error, UnidentifiedImageError) as e: logging.error(f"Error memproses gambar base64: {e}") return jsonify({"error": f"Data base64 tidak valid atau format gambar tidak didukung."}), 400 except Exception as e: logging.error(f"Error tak terduga saat memproses gambar: {e}") return jsonify({"error": "Gagal memproses gambar."}), 500 try: image_rgb = np.array(img_pil) img_tensor = TRANSFORMS(image=image_rgb)['image'].unsqueeze(0).to(DEVICE) with torch.no_grad(): type_logits, object_logits, text_log_probs = MODEL(img_tensor) type_prob = torch.softmax(type_logits, dim=1) type_conf, type_pred_idx = torch.max(type_prob, dim=1) pred_type = MAPPINGS['idx_to_type'].get(str(type_pred_idx.item()), 'Tipe Tidak Dikenal') response = { "predicted_type": pred_type, "type_confidence": f"{type_conf.item():.2%}", "prediction": None, "prediction_confidence": None, "error": None } if pred_type == 'object': obj_prob = torch.softmax(object_logits, dim=1) obj_conf, obj_pred_idx = torch.max(obj_prob, dim=1) pred_obj = MAPPINGS['idx_to_object'].get(str(obj_pred_idx.item()), 'Objek Tidak Dikenal') response["prediction"] = pred_obj response["prediction_confidence"] = f"{obj_conf.item():.2%}" elif pred_type == 'text': pred_text, confidence = ctc_decoder_with_confidence(text_log_probs, MAPPINGS['ctc_idx_to_char'], MAPPINGS['ctc_blank_idx']) response["prediction"] = pred_text response["prediction_confidence"] = f"{confidence:.2%}" logging.info(f"Prediksi berhasil: Tipe='{response['predicted_type']}', Hasil='{response['prediction']}', Conf='{response['prediction_confidence']}'") return jsonify(response), 200 except Exception as e: logging.error(f"Error saat inferensi model: {e}", exc_info=True) return jsonify({"error": "Terjadi kesalahan pada server saat melakukan prediksi."}), 500 # ============================================================================== # BAGIAN 6: MENJALANKAN SERVER (UNTUK HUGGING FACE SPACES) # ============================================================================== # Menggunakan file model baru yang sudah terkuantisasi MODEL_FILE_PATH = "best_model_quantized.pth" MAPPINGS_FILE_PATH = "mappings.json" # Inisialisasi server saat aplikasi dimulai initialize_server(MODEL_FILE_PATH, MAPPINGS_FILE_PATH) # Berguna untuk pengujian lokal if __name__ == '__main__': # Mode debug dimatikan dan port disamakan dengan Dockerfile untuk konsistensi app.run(host='0.0.0.0', port=7860)