File size: 3,693 Bytes
5fc6e5d
0d60ae9
5fc6e5d
66e683e
5fc6e5d
66e683e
5fc6e5d
 
 
 
66e683e
5fc6e5d
 
 
0d60ae9
 
5fc6e5d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66e683e
 
 
5fc6e5d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66e683e
 
 
5fc6e5d
 
 
 
 
 
 
 
 
 
 
66e683e
 
 
5fc6e5d
 
 
 
 
66e683e
 
5fc6e5d
 
66e683e
5fc6e5d
 
 
 
 
 
 
 
 
66e683e
5fc6e5d
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
import base64
import logging
import os
from typing import Literal

from fastapi import FastAPI, HTTPException, Query
from fastapi.responses import JSONResponse
import gradio as gr

from turing.api.demo import create_demo
from turing.api.resource_monitoring import PrometheusBodyMiddleware, instrumentator
from turing.api.schemas import PredictionRequest, PredictionResponse
from turing.modeling.predict import ModelInference

logger = logging.getLogger(__name__)


def get_logo_b64_src(filename="logo_header.svg"):
    """read SVG and convert it into a string Base64 for HTML."""
    try:
        base_path = os.path.dirname(os.path.abspath(__file__))
        target_path = os.path.join(base_path, "..", "..", "reports", "figures", filename)
        target_path = os.path.normpath(target_path)
        
        with open(target_path, "rb") as f:
            encoded = base64.b64encode(f.read()).decode("utf-8")
        return f"data:image/svg+xml;base64,{encoded}"
    except Exception as e:
        print(f"Unable to load logo for API: {e}")
        return "" 


# load logo
logo_src = get_logo_b64_src()

# html
logo_html_big = f"""
<a href="/gradio">
    <img src="{logo_src}" width="150" style="display: block; margin: 10px 0;">
</a>
"""

# description
description_md = f"""
API for classifying code comments.

You can interact with the model directly using the visual interface. 
Click the logo below to open it:

{logo_html_big}

"""

app = FastAPI(
    title="Turing Team Code Classification API",
    description=description_md,
    version="1.0.0"
)

## Add Prometheus middleware
app.add_middleware(PrometheusBodyMiddleware)

@app.get("/manifest.json")
def get_manifest():
    return JSONResponse(content={
        "name": "Turing App",
        "short_name": "Turing",
        "start_url": "/gradio",
        "display": "standalone",
        "background_color": "#ffffff",
        "theme_color": "#000000",
        "icons": []
    })

# Global inference engine instance
inference_engine = ModelInference()
demo = create_demo(inference_engine)

# Instrument the app with Prometheus metrics
instrumentator.instrument(app).expose(app,include_in_schema=False, should_gzip=True)
app = gr.mount_gradio_app(app, demo, path="/gradio")

@app.get("/")
def health_check():
    """
    Root endpoint to verify API status.
    """
    return {"status": "ok", "message": "Turing Code Classification API is ready.", "ui_url": "/gradio"}


@app.post("/predict", response_model=PredictionResponse)
async def predict(request: PredictionRequest, language: Literal["java", "python", "pharo"] = Query(
        ...
    )):
    """
    Endpoint to classify a list of code comments.
    Dynamically loads the model from MLflow based on the request parameters.
    """
    try:
        logger.info(f"Received prediction request for language: {language}")
        
        # Perform prediction using the inference engine
        raw, predictions, run_id, artifact = inference_engine.predict_payload(
            texts=request.texts, language=language
        )

        # Ensure predictions are serializable (convert numpy arrays to lists)
        if hasattr(predictions, "tolist"):
            predictions = predictions.tolist()

        return PredictionResponse(
            predictions=raw.tolist(),
            labels=predictions,
            model_info={"artifact": artifact, "language": language},
        )

    except Exception as e:
        logger.error(f"Prediction failed: {str(e)}")
        raise HTTPException(status_code=500, detail=str(e))


# Entry point for running the API directly with python
if __name__ == "__main__":
    import uvicorn

    uvicorn.run(app, host="127.0.0.1", port=7860)