|
|
|
|
|
import re |
|
|
import json |
|
|
from typing import Any |
|
|
from typing import Dict |
|
|
from typing import List |
|
|
from pathlib import Path |
|
|
from loguru import logger |
|
|
from typing import Optional |
|
|
from datetime import datetime |
|
|
from dataclasses import dataclass |
|
|
from detector.orchestrator import DetectionResult |
|
|
from detector.attribution import AttributionResult |
|
|
from reporter.reasoning_generator import DetailedReasoning |
|
|
from reporter.reasoning_generator import ReasoningGenerator |
|
|
|
|
|
|
|
|
@dataclass |
|
|
class DetailedMetric: |
|
|
""" |
|
|
Metric data structure with sub-metrics |
|
|
""" |
|
|
name : str |
|
|
ai_probability : float |
|
|
human_probability : float |
|
|
confidence : float |
|
|
verdict : str |
|
|
description : str |
|
|
detailed_metrics : Dict[str, float] |
|
|
weight : float |
|
|
|
|
|
|
|
|
class ReportGenerator: |
|
|
""" |
|
|
Generates comprehensive detection reports with detailed metrics |
|
|
|
|
|
Supports: |
|
|
- JSON (structured data with all details) |
|
|
- PDF (printable reports with tables and formatting) |
|
|
""" |
|
|
def __init__(self, output_dir: Optional[Path] = None): |
|
|
""" |
|
|
Initialize report generator |
|
|
|
|
|
Arguments: |
|
|
---------- |
|
|
output_dir { str } : Directory for saving reports (default: data/reports) |
|
|
""" |
|
|
if (output_dir is None): |
|
|
output_dir = Path(__file__).parent.parent / "data" / "reports" |
|
|
|
|
|
self.output_dir = Path(output_dir) |
|
|
self.output_dir.mkdir(parents = True, |
|
|
exist_ok = True, |
|
|
) |
|
|
|
|
|
self.reasoning_generator = ReasoningGenerator() |
|
|
|
|
|
logger.info(f"ReportGenerator initialized (output_dir={self.output_dir})") |
|
|
|
|
|
|
|
|
def generate_complete_report(self, detection_result: DetectionResult, attribution_result: Optional[AttributionResult] = None, highlighted_sentences: Optional[List] = None, |
|
|
formats: List[str] = ["json", "pdf"], filename_prefix: str = "ai_detection_report") -> Dict[str, str]: |
|
|
""" |
|
|
Generate comprehensive report in JSON and PDF formats with detailed metrics |
|
|
|
|
|
Arguments: |
|
|
---------- |
|
|
detection_result : Detection analysis result |
|
|
|
|
|
attribution_result : Model attribution result (optional) |
|
|
|
|
|
highlighted_sentences : List of highlighted sentences (optional) |
|
|
|
|
|
formats : List of formats to generate (json, pdf) |
|
|
|
|
|
filename_prefix : Prefix for output filenames |
|
|
|
|
|
Returns: |
|
|
-------- |
|
|
{ dict } : Dictionary mapping format to filepath |
|
|
""" |
|
|
|
|
|
detection_dict = detection_result.to_dict() if hasattr(detection_result, 'to_dict') else detection_result |
|
|
|
|
|
|
|
|
if ("detection_result" in detection_dict): |
|
|
detection_data = detection_dict["detection_result"] |
|
|
logger.info("Extracted detection_result from outer dict") |
|
|
|
|
|
else: |
|
|
detection_data = detection_dict |
|
|
logger.info("Using detection_dict directly") |
|
|
|
|
|
|
|
|
reasoning = self.reasoning_generator.generate(ensemble_result = detection_result.ensemble_result, |
|
|
metric_results = detection_result.metric_results, |
|
|
domain = detection_result.domain_prediction.primary_domain, |
|
|
attribution_result = attribution_result, |
|
|
text_length = detection_result.processed_text.word_count, |
|
|
) |
|
|
|
|
|
|
|
|
detailed_metrics = self._extract_detailed_metrics(detection_data = detection_data) |
|
|
|
|
|
|
|
|
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") |
|
|
|
|
|
generated_files = dict() |
|
|
|
|
|
|
|
|
if ("json" in formats): |
|
|
json_path = self._generate_json_report(detection_data = detection_data, |
|
|
detection_dict_full = detection_dict, |
|
|
reasoning = reasoning, |
|
|
detailed_metrics = detailed_metrics, |
|
|
attribution_result = attribution_result, |
|
|
highlighted_sentences = highlighted_sentences, |
|
|
filename = f"{filename_prefix}_{timestamp}.json", |
|
|
) |
|
|
generated_files["json"] = str(json_path) |
|
|
|
|
|
if ("pdf" in formats): |
|
|
try: |
|
|
pdf_path = self._generate_pdf_report(detection_data = detection_data, |
|
|
detection_dict_full = detection_dict, |
|
|
reasoning = reasoning, |
|
|
detailed_metrics = detailed_metrics, |
|
|
attribution_result = attribution_result, |
|
|
highlighted_sentences = highlighted_sentences, |
|
|
filename = f"{filename_prefix}_{timestamp}.pdf", |
|
|
) |
|
|
generated_files["pdf"] = str(pdf_path) |
|
|
|
|
|
except Exception as e: |
|
|
logger.warning(f"PDF generation failed: {repr(e)}") |
|
|
logger.info("Install reportlab for PDF support: pip install reportlab") |
|
|
|
|
|
logger.info(f"Generated {len(generated_files)} report(s): {list(generated_files.keys())}") |
|
|
|
|
|
return generated_files |
|
|
|
|
|
|
|
|
def _extract_detailed_metrics(self, detection_data: Dict) -> List[DetailedMetric]: |
|
|
""" |
|
|
Extract detailed metrics with sub-metrics from ACTUAL detection result |
|
|
""" |
|
|
detailed_metrics = list() |
|
|
metrics_data = detection_data.get("metrics", {}) |
|
|
ensemble_data = detection_data.get("ensemble", {}) |
|
|
|
|
|
|
|
|
metric_weights = ensemble_data.get("metric_contributions", {}) |
|
|
|
|
|
|
|
|
for metric_name, metric_result in metrics_data.items(): |
|
|
if (not isinstance(metric_result, dict)): |
|
|
logger.warning(f"Metric {metric_name} is not a dict: {type(metric_result)}") |
|
|
continue |
|
|
|
|
|
if (metric_result.get("error") is not None): |
|
|
logger.warning(f"Metric {metric_name} has error: {metric_result.get('error')}") |
|
|
continue |
|
|
|
|
|
|
|
|
ai_prob = metric_result.get("ai_probability", 0) |
|
|
human_prob = metric_result.get("human_probability", 0) |
|
|
confidence = metric_result.get("confidence", 0) |
|
|
|
|
|
|
|
|
if (human_prob >= 0.6): |
|
|
verdict = "HUMAN" |
|
|
|
|
|
elif (ai_prob >= 0.6): |
|
|
verdict = "AI" |
|
|
|
|
|
elif (ai_prob > 0.4 and ai_prob < 0.6): |
|
|
verdict = "MIXED" |
|
|
|
|
|
elif (human_prob > 0.4 and human_prob < 0.6): |
|
|
verdict = "MIXED" |
|
|
|
|
|
else: |
|
|
|
|
|
if (human_prob > ai_prob): |
|
|
verdict = "HUMAN" |
|
|
|
|
|
elif (ai_prob > human_prob): |
|
|
verdict = "AI" |
|
|
|
|
|
else: |
|
|
verdict = "MIXED" |
|
|
|
|
|
|
|
|
weight = 0.0 |
|
|
if (metric_name in metric_weights): |
|
|
weight = metric_weights[metric_name].get("weight", 0.0) |
|
|
|
|
|
|
|
|
detailed_metrics_data = self._extract_metric_details(metric_name = metric_name, |
|
|
metric_result = metric_result, |
|
|
) |
|
|
|
|
|
|
|
|
description = self._get_metric_description(metric_name = metric_name) |
|
|
|
|
|
detailed_metrics.append(DetailedMetric(name = metric_name, |
|
|
ai_probability = ai_prob * 100, |
|
|
human_probability = human_prob * 100, |
|
|
confidence = confidence * 100, |
|
|
verdict = verdict, |
|
|
description = description, |
|
|
detailed_metrics = detailed_metrics_data, |
|
|
weight = weight * 100, |
|
|
) |
|
|
) |
|
|
|
|
|
logger.info(f"Extracted {len(detailed_metrics)} detailed metrics") |
|
|
|
|
|
return detailed_metrics |
|
|
|
|
|
|
|
|
def _extract_metric_details(self, metric_name: str, metric_result: Dict) -> Dict[str, float]: |
|
|
""" |
|
|
Extract detailed sub-metrics from metric result |
|
|
""" |
|
|
details = dict() |
|
|
|
|
|
|
|
|
if metric_result.get("details"): |
|
|
|
|
|
for key, value in metric_result["details"].items(): |
|
|
if (isinstance(value, (int, float))): |
|
|
|
|
|
if ("perplexity" in key.lower()): |
|
|
details[key] = float(f"{value:.2f}") |
|
|
|
|
|
elif ("entropy" in key.lower()): |
|
|
details[key] = float(f"{value:.2f}") |
|
|
|
|
|
elif (("score" in key.lower()) or ("ratio" in key.lower())): |
|
|
details[key] = float(f"{value:.4f}") |
|
|
|
|
|
elif ("probability" in key.lower()): |
|
|
details[key] = float(f"{value:.4f}") |
|
|
|
|
|
else: |
|
|
details[key] = float(f"{value:.3f}") |
|
|
|
|
|
else: |
|
|
details[key] = value |
|
|
|
|
|
|
|
|
if not details: |
|
|
details = {"ai_probability" : metric_result.get("ai_probability", 0) * 100, |
|
|
"human_probability" : metric_result.get("human_probability", 0) * 100, |
|
|
"confidence" : metric_result.get("confidence", 0) * 100, |
|
|
"score" : metric_result.get("raw_score", 0) * 100, |
|
|
} |
|
|
|
|
|
return details |
|
|
|
|
|
|
|
|
def _get_metric_description(self, metric_name: str) -> str: |
|
|
""" |
|
|
Get description for each metric type |
|
|
""" |
|
|
descriptions = {"structural" : "Analyzes sentence structure, length patterns, and statistical features", |
|
|
"perplexity" : "Measures text predictability using language model cross-entropy", |
|
|
"entropy" : "Evaluates token diversity and sequence unpredictability", |
|
|
"semantic_analysis" : "Examines semantic coherence, topic consistency, and logical flow", |
|
|
"linguistic" : "Assesses grammatical patterns, syntactic complexity, and style markers", |
|
|
"multi_perturbation_stability" : "Tests text stability under perturbation using curvature analysis", |
|
|
} |
|
|
|
|
|
return descriptions.get(metric_name, "Advanced text analysis metric.") |
|
|
|
|
|
|
|
|
def _generate_json_report(self, detection_data: Dict, detection_dict_full: Dict, reasoning: DetailedReasoning, detailed_metrics: List[DetailedMetric], |
|
|
attribution_result: Optional[AttributionResult], highlighted_sentences: Optional[List] = None, filename: str = None) -> Path: |
|
|
""" |
|
|
Generate JSON format report with detailed metrics |
|
|
""" |
|
|
|
|
|
metrics_data = list() |
|
|
|
|
|
for metric in detailed_metrics: |
|
|
metrics_data.append({"name" : metric.name, |
|
|
"ai_probability" : metric.ai_probability, |
|
|
"human_probability" : metric.human_probability, |
|
|
"confidence" : metric.confidence, |
|
|
"verdict" : metric.verdict, |
|
|
"description" : metric.description, |
|
|
"weight" : metric.weight, |
|
|
"detailed_metrics" : metric.detailed_metrics, |
|
|
}) |
|
|
|
|
|
|
|
|
highlighted_data = None |
|
|
|
|
|
if highlighted_sentences: |
|
|
highlighted_data = list() |
|
|
|
|
|
for sent in highlighted_sentences: |
|
|
highlighted_data.append({"text" : sent.text, |
|
|
"ai_probability" : sent.ai_probability, |
|
|
"confidence" : sent.confidence, |
|
|
"color_class" : sent.color_class, |
|
|
"index" : sent.index, |
|
|
}) |
|
|
|
|
|
|
|
|
attribution_data = None |
|
|
|
|
|
if attribution_result: |
|
|
attribution_data = {"predicted_model" : attribution_result.predicted_model.value, |
|
|
"confidence" : attribution_result.confidence, |
|
|
"model_probabilities" : attribution_result.model_probabilities, |
|
|
"reasoning" : attribution_result.reasoning, |
|
|
"fingerprint_matches" : attribution_result.fingerprint_matches, |
|
|
"domain_used" : attribution_result.domain_used.value, |
|
|
"metric_contributions": attribution_result.metric_contributions, |
|
|
} |
|
|
|
|
|
|
|
|
ensemble_data = detection_data.get("ensemble", {}) |
|
|
analysis_data = detection_data.get("analysis", {}) |
|
|
metrics_data_dict = detection_data.get("metrics", {}) |
|
|
performance_data = detection_data.get("performance", {}) |
|
|
|
|
|
report_data = {"report_metadata" : {"generated_at" : datetime.now().isoformat(), |
|
|
"version" : "1.0.0", |
|
|
"format" : "json", |
|
|
"report_id" : filename.replace('.json', ''), |
|
|
}, |
|
|
"overall_results" : {"final_verdict" : ensemble_data.get("final_verdict", "Unknown"), |
|
|
"ai_probability" : ensemble_data.get("ai_probability", 0), |
|
|
"human_probability" : ensemble_data.get("human_probability", 0), |
|
|
"mixed_probability" : ensemble_data.get("mixed_probability", 0), |
|
|
"overall_confidence" : ensemble_data.get("overall_confidence", 0), |
|
|
"uncertainty_score" : ensemble_data.get("uncertainty_score", 0), |
|
|
"consensus_level" : ensemble_data.get("consensus_level", 0), |
|
|
"domain" : analysis_data.get("domain", "general"), |
|
|
"domain_confidence" : analysis_data.get("domain_confidence", 0), |
|
|
"text_length" : analysis_data.get("text_length", 0), |
|
|
"sentence_count" : analysis_data.get("sentence_count", 0), |
|
|
}, |
|
|
"ensemble_analysis" : {"method_used" : "confidence_calibrated", |
|
|
"metric_weights" : ensemble_data.get("metric_contributions", {}), |
|
|
"reasoning" : ensemble_data.get("reasoning", []), |
|
|
}, |
|
|
"detailed_metrics" : metrics_data, |
|
|
"detection_reasoning" : {"summary" : reasoning.summary, |
|
|
"key_indicators" : reasoning.key_indicators, |
|
|
"metric_explanations" : reasoning.metric_explanations, |
|
|
"supporting_evidence" : reasoning.supporting_evidence, |
|
|
"contradicting_evidence" : reasoning.contradicting_evidence, |
|
|
"confidence_explanation" : reasoning.confidence_explanation, |
|
|
"domain_analysis" : reasoning.domain_analysis, |
|
|
"ensemble_analysis" : reasoning.ensemble_analysis, |
|
|
"uncertainty_analysis" : reasoning.uncertainty_analysis, |
|
|
"recommendations" : reasoning.recommendations, |
|
|
}, |
|
|
"highlighted_text" : highlighted_data, |
|
|
"model_attribution" : attribution_data, |
|
|
"performance_metrics" : {"total_processing_time" : performance_data.get("total_time", 0), |
|
|
"metrics_execution_time" : performance_data.get("metrics_time", {}), |
|
|
"warnings" : detection_data.get("warnings", []), |
|
|
"errors" : detection_data.get("errors", []), |
|
|
} |
|
|
} |
|
|
|
|
|
output_path = self.output_dir / filename |
|
|
|
|
|
with open(output_path, 'w', encoding='utf-8') as f: |
|
|
json.dump(obj = report_data, |
|
|
fp = f, |
|
|
indent = 4, |
|
|
ensure_ascii = False, |
|
|
) |
|
|
|
|
|
logger.info(f"JSON report saved: {output_path}") |
|
|
|
|
|
return output_path |
|
|
|
|
|
|
|
|
def _generate_pdf_report(self, detection_data: Dict, detection_dict_full: Dict, reasoning: DetailedReasoning, detailed_metrics: List[DetailedMetric], |
|
|
attribution_result: Optional[AttributionResult], highlighted_sentences: Optional[List] = None, filename: str = None) -> Path: |
|
|
""" |
|
|
Generate PDF format report with detailed metrics |
|
|
""" |
|
|
try: |
|
|
from reportlab.lib import colors |
|
|
from reportlab.lib.units import cm |
|
|
from reportlab.platypus import Table |
|
|
from reportlab.lib.units import inch |
|
|
from reportlab.platypus import Spacer |
|
|
from reportlab.lib.pagesizes import A4 |
|
|
from reportlab.lib.enums import TA_LEFT |
|
|
from reportlab.platypus import PageBreak |
|
|
from reportlab.platypus import Paragraph |
|
|
from reportlab.lib.enums import TA_RIGHT |
|
|
from reportlab.graphics import renderPDF |
|
|
from reportlab.lib.enums import TA_CENTER |
|
|
from reportlab.platypus import TableStyle |
|
|
from reportlab.pdfgen.canvas import Canvas |
|
|
from reportlab.lib.enums import TA_JUSTIFY |
|
|
from reportlab.lib.pagesizes import letter |
|
|
from reportlab.graphics.shapes import Line |
|
|
from reportlab.graphics.shapes import Rect |
|
|
from reportlab.platypus import KeepTogether |
|
|
from reportlab.graphics.shapes import Circle |
|
|
from reportlab.graphics.shapes import Drawing |
|
|
from reportlab.lib.styles import ParagraphStyle |
|
|
from reportlab.platypus import SimpleDocTemplate |
|
|
from reportlab.graphics.charts.piecharts import Pie |
|
|
from reportlab.platypus.flowables import HRFlowable |
|
|
from reportlab.lib.styles import getSampleStyleSheet |
|
|
from reportlab.graphics.charts.textlabels import Label |
|
|
from reportlab.graphics.widgets.markers import makeMarker |
|
|
|
|
|
except ImportError: |
|
|
raise ImportError("reportlab is required for PDF generation. Install: pip install reportlab") |
|
|
|
|
|
output_path = self.output_dir / filename |
|
|
|
|
|
|
|
|
doc = SimpleDocTemplate(str(output_path), |
|
|
pagesize = A4, |
|
|
rightMargin = 0.75*inch, |
|
|
leftMargin = 0.75*inch, |
|
|
topMargin = 0.75*inch, |
|
|
bottomMargin = 0.75*inch, |
|
|
) |
|
|
|
|
|
|
|
|
elements = list() |
|
|
styles = getSampleStyleSheet() |
|
|
|
|
|
|
|
|
PRIMARY_COLOR = '#3b82f6' |
|
|
SUCCESS_COLOR = '#10b981' |
|
|
WARNING_COLOR = '#f59e0b' |
|
|
DANGER_COLOR = '#ef4444' |
|
|
INFO_COLOR = '#8b5cf6' |
|
|
GRAY_LIGHT = '#f8fafc' |
|
|
GRAY_MEDIUM = '#e2e8f0' |
|
|
GRAY_DARK = '#334155' |
|
|
TEXT_COLOR = '#1e293b' |
|
|
|
|
|
|
|
|
title_style = ParagraphStyle('PremiumTitle', |
|
|
parent = styles['Heading1'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 28, |
|
|
textColor = PRIMARY_COLOR, |
|
|
spaceAfter = 20, |
|
|
alignment = TA_CENTER, |
|
|
) |
|
|
|
|
|
subtitle_style = ParagraphStyle('PremiumSubtitle', |
|
|
parent = styles['Normal'], |
|
|
fontName = 'Helvetica', |
|
|
fontSize = 12, |
|
|
textColor = GRAY_DARK, |
|
|
spaceAfter = 30, |
|
|
alignment = TA_CENTER, |
|
|
) |
|
|
|
|
|
filename_style = ParagraphStyle('FilenameStyle', |
|
|
parent = styles['Normal'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 10, |
|
|
textColor = GRAY_DARK, |
|
|
spaceAfter = 10, |
|
|
alignment = TA_CENTER, |
|
|
) |
|
|
|
|
|
section_style = ParagraphStyle('PremiumSection', |
|
|
parent = styles['Heading2'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 18, |
|
|
textColor = TEXT_COLOR, |
|
|
spaceAfter = 12, |
|
|
spaceBefore = 20, |
|
|
underlineWidth = 1, |
|
|
underlineColor = PRIMARY_COLOR, |
|
|
) |
|
|
|
|
|
subsection_style = ParagraphStyle('PremiumSubSection', |
|
|
parent = styles['Heading3'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 14, |
|
|
textColor = GRAY_DARK, |
|
|
spaceAfter = 8, |
|
|
spaceBefore = 16, |
|
|
) |
|
|
|
|
|
key_indicators_style = ParagraphStyle('KeyIndicatorsStyle', |
|
|
parent = styles['Heading2'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 18, |
|
|
textColor = TEXT_COLOR, |
|
|
spaceAfter = 12, |
|
|
spaceBefore = 20, |
|
|
underlineWidth = 1, |
|
|
underlineColor = PRIMARY_COLOR, |
|
|
) |
|
|
|
|
|
body_style = ParagraphStyle('PremiumBody', |
|
|
parent = styles['BodyText'], |
|
|
fontName = 'Helvetica', |
|
|
fontSize = 11, |
|
|
textColor = TEXT_COLOR, |
|
|
alignment = TA_JUSTIFY, |
|
|
spaceAfter = 8, |
|
|
) |
|
|
|
|
|
|
|
|
page2_body_style = ParagraphStyle('Page2Body', |
|
|
parent = styles['BodyText'], |
|
|
fontName = 'Helvetica', |
|
|
fontSize = 11, |
|
|
textColor = TEXT_COLOR, |
|
|
alignment = TA_JUSTIFY, |
|
|
spaceAfter = 8, |
|
|
) |
|
|
|
|
|
bullet_style = ParagraphStyle('BulletStyle', |
|
|
parent = styles['BodyText'], |
|
|
fontName = 'Helvetica', |
|
|
fontSize = 11, |
|
|
textColor = TEXT_COLOR, |
|
|
alignment = TA_LEFT, |
|
|
spaceAfter = 6, |
|
|
leftIndent = 20, |
|
|
) |
|
|
|
|
|
bold_style = ParagraphStyle('BoldStyle', |
|
|
parent = styles['BodyText'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 11, |
|
|
textColor = TEXT_COLOR, |
|
|
alignment = TA_LEFT, |
|
|
spaceAfter = 8, |
|
|
) |
|
|
|
|
|
small_bold_style = ParagraphStyle('SmallBoldStyle', |
|
|
parent = styles['BodyText'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 9, |
|
|
textColor = TEXT_COLOR, |
|
|
alignment = TA_LEFT, |
|
|
spaceAfter = 4, |
|
|
) |
|
|
|
|
|
small_style = ParagraphStyle('SmallStyle', |
|
|
parent = styles['BodyText'], |
|
|
fontName = 'Helvetica', |
|
|
fontSize = 9, |
|
|
textColor = TEXT_COLOR, |
|
|
alignment = TA_LEFT, |
|
|
spaceAfter = 4, |
|
|
) |
|
|
|
|
|
footer_style = ParagraphStyle('FooterStyle', |
|
|
parent = styles['Normal'], |
|
|
fontName = 'Helvetica', |
|
|
fontSize = 9, |
|
|
textColor = GRAY_DARK, |
|
|
alignment = TA_CENTER, |
|
|
) |
|
|
|
|
|
print (detection_dict_full.keys()) |
|
|
|
|
|
|
|
|
ensemble_data = detection_data.get("ensemble", {}) |
|
|
analysis_data = detection_data.get("analysis", {}) |
|
|
performance_data = detection_data.get("performance", {}) |
|
|
|
|
|
|
|
|
file_info = detection_data.get("file_info", {}) |
|
|
|
|
|
|
|
|
original_filename = file_info.get("filename", "Unknown") |
|
|
|
|
|
|
|
|
ai_prob = ensemble_data.get("ai_probability", 0) * 100 |
|
|
human_prob = ensemble_data.get("human_probability", 0) * 100 |
|
|
mixed_prob = ensemble_data.get("mixed_probability", 0) * 100 |
|
|
confidence = ensemble_data.get("overall_confidence", 0) * 100 |
|
|
uncertainty = ensemble_data.get("uncertainty_score", 0) * 100 |
|
|
consensus = ensemble_data.get("consensus_level", 0) * 100 |
|
|
final_verdict = ensemble_data.get("final_verdict", "Unknown") |
|
|
total_time = performance_data.get("total_time", 0) |
|
|
|
|
|
|
|
|
if ("Human".lower() in final_verdict.lower()): |
|
|
verdict_color = SUCCESS_COLOR |
|
|
|
|
|
elif ("AI".lower() in final_verdict.lower()): |
|
|
verdict_color = DANGER_COLOR |
|
|
|
|
|
elif ("Mixed".lower() in final_verdict.lower()): |
|
|
verdict_color = WARNING_COLOR |
|
|
|
|
|
else: |
|
|
verdict_color = PRIMARY_COLOR |
|
|
|
|
|
|
|
|
|
|
|
header_style = ParagraphStyle('HeaderStyle', |
|
|
parent = styles['Normal'], |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 10, |
|
|
textColor = GRAY_DARK, |
|
|
alignment = TA_RIGHT, |
|
|
) |
|
|
|
|
|
elements.append(Paragraph("AI DETECTION ANALYTICS", header_style)) |
|
|
|
|
|
elements.append(HRFlowable(width = "100%", |
|
|
thickness = 1, |
|
|
color = PRIMARY_COLOR, |
|
|
spaceAfter = 15, |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("AI Text Detection Analysis Report", title_style)) |
|
|
elements.append(Paragraph(f"Generated on {datetime.now().strftime('%B %d, %Y at %I:%M %p')}", subtitle_style)) |
|
|
|
|
|
|
|
|
elements.append(Paragraph(f"Analyzed File: {original_filename}", filename_style)) |
|
|
elements.append(Spacer(1, 0.1*inch)) |
|
|
|
|
|
|
|
|
elements.append(HRFlowable(width = "80%", |
|
|
thickness = 2, |
|
|
color = PRIMARY_COLOR, |
|
|
spaceBefore = 10, |
|
|
spaceAfter = 25, |
|
|
hAlign = 'CENTER', |
|
|
) |
|
|
) |
|
|
|
|
|
|
|
|
stats_data = [['Text Source', 'AI', 'HUMAN', 'MIXED'], |
|
|
['Probability', f"{ai_prob:.1f}%", f"{human_prob:.1f}%", f"{mixed_prob:.1f}%"] |
|
|
] |
|
|
|
|
|
stats_table = Table(stats_data, colWidths = [1.5*inch, 1*inch, 1*inch, 1*inch]) |
|
|
|
|
|
stats_table.setStyle(TableStyle([('BACKGROUND', (0, 0), (-1, 0), PRIMARY_COLOR), |
|
|
('TEXTCOLOR', (0, 0), (-1, 0), colors.white), |
|
|
('BACKGROUND', (1, 1), (1, 1), DANGER_COLOR), |
|
|
('BACKGROUND', (2, 1), (2, 1), SUCCESS_COLOR), |
|
|
('BACKGROUND', (3, 1), (3, 1), WARNING_COLOR), |
|
|
('TEXTCOLOR', (1, 1), (-1, 1), colors.white), |
|
|
('ALIGN', (0, 0), (-1, -1), 'CENTER'), |
|
|
('FONTNAME', (0, 0), (-1, 0), 'Helvetica-Bold'), |
|
|
('FONTSIZE', (0, 0), (-1, -1), 11), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 8), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 8), |
|
|
('GRID', (0, 0), (-1, -1), 0.5, colors.white), |
|
|
('BOX', (0, 0), (-1, -1), 1, PRIMARY_COLOR), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(stats_table) |
|
|
elements.append(Spacer(1, 0.3*inch)) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("DETECTION VERDICT", section_style)) |
|
|
|
|
|
verdict_box_data = [[Paragraph(f"<font size=18 color='{verdict_color}'><b>{final_verdict.upper()}</b></font>", ParagraphStyle('VerdictText', alignment=TA_CENTER)), |
|
|
Paragraph(f"<font size=12>Confidence: <b>{confidence:.1f}%</b></font><br/>" |
|
|
f"<font size=10>Uncertainty: {uncertainty:.1f}% | Consensus: {consensus:.1f}%</font>", |
|
|
ParagraphStyle('VerdictDetails', alignment=TA_CENTER)) |
|
|
]] |
|
|
|
|
|
verdict_box = Table(verdict_box_data, colWidths = [2.5*inch, 3*inch]) |
|
|
|
|
|
verdict_box.setStyle(TableStyle([('BACKGROUND', (0, 0), (0, 0), GRAY_LIGHT), |
|
|
('BACKGROUND', (1, 0), (1, 0), GRAY_LIGHT), |
|
|
('BOX', (0, 0), (-1, -1), 1, verdict_color), |
|
|
('ROUNDEDCORNERS', [10, 10, 10, 10]), |
|
|
('ALIGN', (0, 0), (-1, -1), 'CENTER'), |
|
|
('VALIGN', (0, 0), (-1, -1), 'MIDDLE'), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 15), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 15), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(verdict_box) |
|
|
elements.append(Spacer(1, 0.3*inch)) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("DETECTION REASONING", section_style)) |
|
|
|
|
|
|
|
|
summary_text = reasoning.summary if hasattr(reasoning, 'summary') else "No reasoning summary available." |
|
|
|
|
|
|
|
|
summary_text = ' '.join(summary_text.split()) |
|
|
|
|
|
|
|
|
summary_text = re.sub(r'\*\*(.*?)\*\*', r'<b>\1</b>', summary_text) |
|
|
|
|
|
|
|
|
sentences = re.split(r'(?<=[.!?])\s+', summary_text) |
|
|
|
|
|
|
|
|
for i, sentence in enumerate(sentences): |
|
|
if sentence.strip(): |
|
|
|
|
|
elements.append(Paragraph(f"<font color='{PRIMARY_COLOR}'>•</font> {sentence.strip()}", bullet_style)) |
|
|
|
|
|
|
|
|
if (i < len(sentences) - 1): |
|
|
|
|
|
elements.append(Spacer(1, 0.08*inch)) |
|
|
|
|
|
|
|
|
if ((hasattr(reasoning, 'key_indicators')) and reasoning.key_indicators and (len(reasoning.key_indicators) > 0)): |
|
|
elements.append(Paragraph("KEY INDICATORS", key_indicators_style)) |
|
|
|
|
|
for indicator in reasoning.key_indicators: |
|
|
if isinstance(indicator, str): |
|
|
|
|
|
indicator = ' '.join(indicator.split()) |
|
|
|
|
|
|
|
|
formatted_indicator = re.sub(r'\*\*(.*?)\*\*', r'<b>\1</b>', indicator) |
|
|
|
|
|
|
|
|
formatted_indicator = formatted_indicator.replace('_', ' ') |
|
|
|
|
|
elements.append(Paragraph(f"<font color='{SUCCESS_COLOR}'>•</font> {formatted_indicator}", body_style)) |
|
|
elements.append(Spacer(1, 0.05*inch)) |
|
|
|
|
|
elements.append(PageBreak()) |
|
|
|
|
|
|
|
|
|
|
|
elements.append(Paragraph("CONTENT ANALYSIS", section_style)) |
|
|
|
|
|
domain = analysis_data.get("domain", "general").replace('_', ' ').upper() |
|
|
|
|
|
|
|
|
domain_confidence = analysis_data.get("domain_confidence", 0) * 100 |
|
|
text_length = analysis_data.get("text_length", 0) |
|
|
sentence_count = analysis_data.get("sentence_count", 0) |
|
|
|
|
|
|
|
|
content_data = [[Paragraph("<b>Content Domain</b>", bold_style), Paragraph(f"<font color='{INFO_COLOR}'><b>{domain}</b></font> ({domain_confidence:.1f}% confidence)", body_style)], |
|
|
[Paragraph("<b>Text Statistics</b>", bold_style), Paragraph(f"{text_length:,} words | {sentence_count:,} sentences", body_style)], |
|
|
[Paragraph("<b>Processing Time</b>", bold_style), Paragraph(f"{total_time:.2f} seconds", body_style)], |
|
|
[Paragraph("<b>Analysis Method</b>", bold_style), Paragraph("Confidence-Weighted Ensemble Aggregation", body_style)], |
|
|
] |
|
|
|
|
|
content_table = Table(content_data, colWidths = [2*inch, 4.5*inch]) |
|
|
|
|
|
content_table.setStyle(TableStyle([('FONTNAME', (0, 0), (0, -1), 'Helvetica-Bold'), |
|
|
('FONTNAME', (1, 0), (1, -1), 'Helvetica'), |
|
|
('FONTSIZE', (0, 0), (-1, -1), 11), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 10), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 10), |
|
|
('GRID', (0, 0), (-1, -1), 0.25, GRAY_MEDIUM), |
|
|
('BACKGROUND', (0, 0), (0, -1), GRAY_LIGHT), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(content_table) |
|
|
elements.append(Spacer(1, 0.4*inch)) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("METRIC CONTRIBUTIONS", section_style)) |
|
|
|
|
|
metric_contributions = ensemble_data.get("metric_contributions", {}) |
|
|
|
|
|
if (metric_contributions and (len(metric_contributions) > 0)): |
|
|
|
|
|
weight_data = [['METRIC NAME', 'ENSEMBLE WEIGHT (%)']] |
|
|
|
|
|
for metric_name, contribution in metric_contributions.items(): |
|
|
weight = contribution.get("weight", 0) * 100 |
|
|
display_name = metric_name.replace('_', ' ').title() |
|
|
|
|
|
weight_data.append([Paragraph(display_name, bold_style), Paragraph(f"{weight:.1f}%", body_style)]) |
|
|
|
|
|
|
|
|
weight_table = Table(weight_data, colWidths = [4*inch, 2.5*inch]) |
|
|
|
|
|
weight_table.setStyle(TableStyle([('BACKGROUND', (0, 0), (-1, 0), PRIMARY_COLOR), |
|
|
('TEXTCOLOR', (0, 0), (-1, 0), colors.white), |
|
|
('ALIGN', (0, 0), (-1, -1), 'LEFT'), |
|
|
('ALIGN', (1, 0), (1, -1), 'RIGHT'), |
|
|
('FONTNAME', (0, 0), (-1, 0), 'Helvetica-Bold'), |
|
|
('FONTSIZE', (0, 0), (-1, -1), 11), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 10), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 10), |
|
|
('GRID', (0, 0), (-1, -1), 0.5, GRAY_MEDIUM), |
|
|
('BACKGROUND', (1, 1), (1, -1), GRAY_LIGHT), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(weight_table) |
|
|
|
|
|
|
|
|
elements.append(Spacer(1, 0.4*inch)) |
|
|
elements.append(HRFlowable(width = "100%", thickness = 1, color = PRIMARY_COLOR, spaceBefore = 10, spaceAfter = 10)) |
|
|
elements.append(Paragraph("<i>Report continues with detailed metric analysis on the following pages...</i>", |
|
|
ParagraphStyle('ContinueStyle', parent = body_style, fontSize = 10, textColor = GRAY_DARK, alignment = TA_CENTER))) |
|
|
|
|
|
elements.append(PageBreak()) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("DETAILED METRIC ANALYSIS", section_style)) |
|
|
elements.append(Spacer(1, 0.2*inch)) |
|
|
|
|
|
|
|
|
page3_metrics = [m for m in detailed_metrics if m.name in ['structural', 'entropy']] |
|
|
|
|
|
for metric in page3_metrics: |
|
|
self._add_detailed_metric_section(elements = elements, |
|
|
metric = metric, |
|
|
small_bold_style = small_bold_style, |
|
|
small_style = small_style, |
|
|
bold_style = bold_style, |
|
|
PRIMARY_COLOR = PRIMARY_COLOR, |
|
|
SUCCESS_COLOR = SUCCESS_COLOR, |
|
|
DANGER_COLOR = DANGER_COLOR, |
|
|
WARNING_COLOR = WARNING_COLOR, |
|
|
GRAY_LIGHT = GRAY_LIGHT, |
|
|
) |
|
|
|
|
|
elements.append(Spacer(1, 0.1*inch)) |
|
|
|
|
|
elements.append(HRFlowable(width = "100%", thickness = 0.5, color = GRAY_MEDIUM, spaceBefore = 5, spaceAfter = 15)) |
|
|
|
|
|
elements.append(PageBreak()) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("DETAILED METRIC ANALYSIS", section_style)) |
|
|
elements.append(Spacer(1, 0.2*inch)) |
|
|
|
|
|
|
|
|
page4_metrics = [m for m in detailed_metrics if m.name in ['perplexity', 'semantic_analysis']] |
|
|
|
|
|
for metric in page4_metrics: |
|
|
self._add_detailed_metric_section(elements = elements, |
|
|
metric = metric, |
|
|
small_bold_style = small_bold_style, |
|
|
small_style = small_style, |
|
|
bold_style = bold_style, |
|
|
PRIMARY_COLOR = PRIMARY_COLOR, |
|
|
SUCCESS_COLOR = SUCCESS_COLOR, |
|
|
DANGER_COLOR = DANGER_COLOR, |
|
|
WARNING_COLOR = WARNING_COLOR, |
|
|
GRAY_LIGHT = GRAY_LIGHT, |
|
|
) |
|
|
|
|
|
elements.append(Spacer(1, 0.3*inch)) |
|
|
elements.append(HRFlowable(width = "100%", thickness = 0.5, color = GRAY_MEDIUM, spaceBefore = 5, spaceAfter = 15)) |
|
|
|
|
|
elements.append(PageBreak()) |
|
|
|
|
|
|
|
|
elements.append(Paragraph("DETAILED METRIC ANALYSIS", section_style)) |
|
|
elements.append(Spacer(1, 0.1*inch)) |
|
|
|
|
|
|
|
|
page5_metrics = [m for m in detailed_metrics if m.name in ['linguistic', 'multi_perturbation_stability']] |
|
|
|
|
|
|
|
|
page5_elements = list() |
|
|
|
|
|
for i, metric in enumerate(page5_metrics): |
|
|
|
|
|
metric_elements = list() |
|
|
|
|
|
|
|
|
self._add_detailed_metric_section(elements = metric_elements, |
|
|
metric = metric, |
|
|
small_bold_style = small_bold_style, |
|
|
small_style = small_style, |
|
|
bold_style = bold_style, |
|
|
PRIMARY_COLOR = PRIMARY_COLOR, |
|
|
SUCCESS_COLOR = SUCCESS_COLOR, |
|
|
DANGER_COLOR = DANGER_COLOR, |
|
|
WARNING_COLOR = WARNING_COLOR, |
|
|
GRAY_LIGHT = GRAY_LIGHT, |
|
|
) |
|
|
|
|
|
|
|
|
page5_elements.extend(metric_elements) |
|
|
|
|
|
|
|
|
if (i < len(page5_metrics) - 1): |
|
|
page5_elements.append(Spacer(1, 0.05*inch)) |
|
|
page5_elements.append(HRFlowable(width = "100%", thickness = 0.5, color = GRAY_MEDIUM, spaceBefore = 5, spaceAfter = 10)) |
|
|
|
|
|
|
|
|
elements.extend(page5_elements) |
|
|
|
|
|
elements.append(PageBreak()) |
|
|
|
|
|
|
|
|
|
|
|
if attribution_result: |
|
|
elements.append(Paragraph("AI MODEL ATTRIBUTION", section_style)) |
|
|
elements.append(Spacer(1, 0.1*inch)) |
|
|
|
|
|
predicted_model = getattr(attribution_result.predicted_model, 'value', str(attribution_result.predicted_model)) |
|
|
predicted_model = predicted_model.replace("_", " ").title() |
|
|
attribution_confidence = getattr(attribution_result, 'confidence', 0) * 100 |
|
|
domain_used = getattr(attribution_result.domain_used, 'value', 'Unknown').upper() |
|
|
|
|
|
|
|
|
attribution_data = [[Paragraph("<b>Predicted Model</b>", bold_style), Paragraph(f"<font color='{INFO_COLOR}'><b>{predicted_model}</b></font>", bold_style)], |
|
|
[Paragraph("<b>Attribution Confidence</b>", bold_style), Paragraph(f"<b>{attribution_confidence:.1f}%</b>", bold_style)], |
|
|
[Paragraph("<b>Domain Used</b>", bold_style), Paragraph(f"<b>{domain_used}</b>", bold_style)] |
|
|
] |
|
|
|
|
|
attribution_table = Table(attribution_data, colWidths = [2.5*inch, 4*inch]) |
|
|
|
|
|
attribution_table.setStyle(TableStyle([('BACKGROUND', (0, 0), (0, -1), GRAY_LIGHT), |
|
|
('FONTNAME', (0, 0), (0, -1), 'Helvetica-Bold'), |
|
|
('FONTSIZE', (0, 0), (-1, -1), 11), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 8), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 8), |
|
|
('GRID', (0, 0), (-1, -1), 0.5, GRAY_MEDIUM), |
|
|
('VALIGN', (0, 0), (-1, -1), 'MIDDLE'), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(attribution_table) |
|
|
elements.append(Spacer(1, 0.2*inch)) |
|
|
|
|
|
|
|
|
model_probs = getattr(attribution_result, 'model_probabilities', {}) |
|
|
if (model_probs and (len(model_probs) > 0)): |
|
|
elements.append(Paragraph("MODEL PROBABILITY DISTRIBUTION", subsection_style)) |
|
|
elements.append(Spacer(1, 0.05*inch)) |
|
|
|
|
|
|
|
|
sorted_models = sorted(model_probs.items(), key = lambda x: x[1], reverse = True)[:10] |
|
|
|
|
|
prob_data = [['LANGUAGE MODEL NAME', 'ATTRIBUTION PROBABILITY']] |
|
|
|
|
|
for model_name, probability in sorted_models: |
|
|
display_name = model_name.replace("_", " ").replace("-", " ").title() |
|
|
prob_data.append([Paragraph(display_name, bold_style), Paragraph(f"{probability:.1%}", bold_style)]) |
|
|
|
|
|
|
|
|
prob_table = Table(prob_data, colWidths = [4*inch, 2.5*inch]) |
|
|
|
|
|
prob_table.setStyle(TableStyle([('BACKGROUND', (0, 0), (-1, 0), INFO_COLOR), |
|
|
('TEXTCOLOR', (0, 0), (-1, 0), colors.white), |
|
|
('ALIGN', (0, 0), (-1, -1), 'LEFT'), |
|
|
('ALIGN', (1, 0), (1, -1), 'RIGHT'), |
|
|
('FONTNAME', (0, 0), (-1, 0), 'Helvetica-Bold'), |
|
|
('FONTSIZE', (0, 0), (-1, -1), 11), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 6), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 6), |
|
|
('GRID', (0, 0), (-1, -1), 0.5, GRAY_MEDIUM), |
|
|
('BACKGROUND', (1, 1), (1, -1), GRAY_LIGHT), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(prob_table) |
|
|
elements.append(Spacer(1, 0.3*inch)) |
|
|
|
|
|
|
|
|
if ((hasattr(reasoning, 'recommendations')) and reasoning.recommendations): |
|
|
elements.append(Paragraph("RECOMMENDATIONS", section_style)) |
|
|
elements.append(Spacer(1, 0.1*inch)) |
|
|
|
|
|
for i, recommendation in enumerate(reasoning.recommendations): |
|
|
|
|
|
if (i % 3 == 0): |
|
|
rec_color = SUCCESS_COLOR |
|
|
|
|
|
elif (i % 3 == 1): |
|
|
rec_color = INFO_COLOR |
|
|
|
|
|
else: |
|
|
rec_color = WARNING_COLOR |
|
|
|
|
|
|
|
|
clean_rec = ' '.join(recommendation.split()) |
|
|
clean_rec = re.sub(r'\*\*(.*?)\*\*', r'<b>\1</b>', clean_rec) |
|
|
clean_rec = clean_rec.replace('_', ' ') |
|
|
|
|
|
rec_box_data = [[Paragraph(f"<font color='{rec_color}'>✓</font> {clean_rec}", body_style)]] |
|
|
rec_box = Table(rec_box_data, colWidths = [6.5*inch]) |
|
|
|
|
|
rec_box.setStyle(TableStyle([('BACKGROUND', (0, 0), (-1, -1), GRAY_LIGHT), |
|
|
('BOX', (0, 0), (-1, -1), 1, rec_color), |
|
|
('PADDING', (0, 0), (-1, -1), 10), |
|
|
('LEFTPADDING', (0, 0), (-1, -1), 8), |
|
|
('BOTTOMMARGIN', (0, 0), (-1, -1), 6), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(rec_box) |
|
|
elements.append(Spacer(1, 0.2*inch)) |
|
|
|
|
|
|
|
|
elements.append(Spacer(1, 0.2*inch)) |
|
|
elements.append(HRFlowable(width = "100%", thickness = 0.5, color = GRAY_MEDIUM, spaceAfter = 8)) |
|
|
|
|
|
|
|
|
report_id = filename.replace('.pdf', '') |
|
|
|
|
|
footer_text = (f"Generated by AI Text Detector v1.0 | " |
|
|
f"Processing Time: {total_time:.2f}s | " |
|
|
f"Report ID: {report_id}") |
|
|
|
|
|
elements.append(Paragraph(footer_text, footer_style)) |
|
|
elements.append(Paragraph("Confidential Analysis Report • © 2025 AI Detection Analytics", |
|
|
ParagraphStyle('Copyright', parent = footer_style, fontSize = 8, textColor = GRAY_MEDIUM))) |
|
|
|
|
|
|
|
|
doc.build(elements) |
|
|
|
|
|
logger.info(f"PDF report saved: {output_path}") |
|
|
|
|
|
return output_path |
|
|
|
|
|
|
|
|
def _add_detailed_metric_section(self, elements, metric, small_bold_style, small_style, bold_style, PRIMARY_COLOR, SUCCESS_COLOR, DANGER_COLOR, WARNING_COLOR, GRAY_LIGHT): |
|
|
""" |
|
|
Add a detailed metric section to the PDF |
|
|
""" |
|
|
|
|
|
from reportlab.platypus import Paragraph, Table, Spacer |
|
|
from reportlab.platypus import TableStyle |
|
|
from reportlab.lib import colors |
|
|
from reportlab.lib.units import inch |
|
|
from reportlab.lib.styles import ParagraphStyle |
|
|
from reportlab.lib.enums import TA_LEFT |
|
|
|
|
|
|
|
|
if (metric.verdict == "HUMAN"): |
|
|
metric_color = SUCCESS_COLOR |
|
|
prob_color = SUCCESS_COLOR |
|
|
|
|
|
elif (metric.verdict == "AI"): |
|
|
metric_color = DANGER_COLOR |
|
|
prob_color = DANGER_COLOR |
|
|
|
|
|
else: |
|
|
metric_color = WARNING_COLOR |
|
|
prob_color = WARNING_COLOR |
|
|
|
|
|
|
|
|
metric_display_name = metric.name.replace('_', ' ').upper() |
|
|
|
|
|
|
|
|
subsection_style = ParagraphStyle('SubsectionStyle', |
|
|
parent = ParagraphStyle('Normal'), |
|
|
fontName = 'Helvetica-Bold', |
|
|
fontSize = 14, |
|
|
textColor = PRIMARY_COLOR, |
|
|
spaceAfter = 8, |
|
|
spaceBefore = 16, |
|
|
alignment=TA_LEFT, |
|
|
) |
|
|
|
|
|
elements.append(Paragraph(f"<b>{metric_display_name}</b>", subsection_style)) |
|
|
elements.append(Paragraph(f"<i>{metric.description}</i>", small_style)) |
|
|
elements.append(Spacer(1, 0.1*inch)) |
|
|
|
|
|
|
|
|
key_metrics_data = [[Paragraph("<b>Verdict</b>", bold_style), Paragraph(f"<font color='{metric_color}'><b>{metric.verdict}</b></font>", bold_style), Paragraph("<b>Weight</b>", bold_style), Paragraph(f"<b>{metric.weight:.1f}%</b>", bold_style)], |
|
|
[Paragraph("<b>AI Probability</b>", bold_style), Paragraph(f"<font color='{prob_color}'><b>{metric.ai_probability:.1f}%</b></font>", bold_style), Paragraph("<b>Confidence</b>", bold_style), Paragraph(f"<b>{metric.confidence:.1f}%</b>", bold_style)] |
|
|
] |
|
|
|
|
|
key_metrics_table = Table(key_metrics_data, colWidths = [1.5*inch, 1.5*inch, 1.5*inch, 1.5*inch]) |
|
|
|
|
|
key_metrics_table.setStyle(TableStyle([('BACKGROUND', (0, 0), (-1, -1), GRAY_LIGHT), |
|
|
('GRID', (0, 0), (-1, -1), 0.5, colors.white), |
|
|
('VALIGN', (0, 0), (-1, -1), 'MIDDLE'), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 8), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 8), |
|
|
('ALIGN', (0, 0), (-1, -1), 'CENTER'), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(key_metrics_table) |
|
|
elements.append(Spacer(1, 0.2*inch)) |
|
|
|
|
|
|
|
|
if metric.detailed_metrics and len(metric.detailed_metrics) > 0: |
|
|
|
|
|
detailed_data = [] |
|
|
|
|
|
|
|
|
sorted_items = sorted(metric.detailed_metrics.items()) |
|
|
|
|
|
|
|
|
for i in range(0, len(sorted_items), 3): |
|
|
row = [] |
|
|
|
|
|
for j in range(3): |
|
|
if i + j < len(sorted_items): |
|
|
key, value = sorted_items[i + j] |
|
|
|
|
|
display_key = key.replace('_', ' ').title() |
|
|
formatted_value = self._format_metric_value(key, value) |
|
|
row.append(Paragraph(f"<font size=9><b>{display_key}:</b></font>", small_bold_style)) |
|
|
row.append(Paragraph(f"<font size=9>{formatted_value}</font>", small_style)) |
|
|
|
|
|
else: |
|
|
row.append("") |
|
|
row.append("") |
|
|
|
|
|
detailed_data.append(row) |
|
|
|
|
|
if detailed_data: |
|
|
|
|
|
col_width = 6.5 * inch / 6 |
|
|
col_widths = [col_width] * 6 |
|
|
|
|
|
detailed_table = Table(detailed_data, colWidths = col_widths) |
|
|
|
|
|
detailed_table.setStyle(TableStyle([('FONTSIZE', (0, 0), (-1, -1), 8), |
|
|
('BOTTOMPADDING', (0, 0), (-1, -1), 3), |
|
|
('TOPPADDING', (0, 0), (-1, -1), 3), |
|
|
('GRID', (0, 0), (-1, -1), 0.2, colors.grey), |
|
|
('VALIGN', (0, 0), (-1, -1), 'MIDDLE'), |
|
|
('ALIGN', (1, 0), (1, -1), 'RIGHT'), |
|
|
('ALIGN', (3, 0), (3, -1), 'RIGHT'), |
|
|
('ALIGN', (5, 0), (5, -1), 'RIGHT'), |
|
|
]) |
|
|
) |
|
|
|
|
|
elements.append(detailed_table) |
|
|
|
|
|
|
|
|
def _format_metric_value(self, key: str, value: Any) -> str: |
|
|
""" |
|
|
Format metric value based on its type |
|
|
""" |
|
|
if not isinstance(value, (int, float)): |
|
|
return str(value) |
|
|
|
|
|
key_lower = key.lower() |
|
|
|
|
|
if ('perplexity' in key_lower): |
|
|
if (value > 1000): |
|
|
return f"{value:,.0f}" |
|
|
|
|
|
else: |
|
|
return f"{value:.2f}" |
|
|
|
|
|
elif (('probability' in key_lower) or ('confidence' in key_lower)): |
|
|
return f"{value:.1f}%" |
|
|
|
|
|
elif ('entropy' in key_lower): |
|
|
return f"{value:.2f}" |
|
|
|
|
|
elif (('ratio' in key_lower) or ('score' in key_lower)): |
|
|
if (0 <= value <= 1): |
|
|
return f"{value:.3f}" |
|
|
|
|
|
else: |
|
|
return f"{value:.2f}" |
|
|
|
|
|
elif (key_lower in ['num_sentences', 'num_words', 'vocabulary_size']): |
|
|
return f"{int(value):,}" |
|
|
|
|
|
elif (('length' in key_lower) or ('size' in key_lower)): |
|
|
return f"{value:.2f}" |
|
|
|
|
|
else: |
|
|
return f"{value:.3f}" |
|
|
|
|
|
|
|
|
|
|
|
__all__ = ["ReportGenerator", |
|
|
"DetailedMetric", |
|
|
] |