from flask import Flask, request, jsonify import torch from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline app = Flask(__name__) # モデルロード(起動時1回) torch.random.manual_seed(0) model = AutoModelForCausalLM.from_pretrained( "microsoft/Phi-3-mini-4k-instruct", device_map="cpu", torch_dtype="auto", trust_remote_code=True ) tokenizer = AutoTokenizer.from_pretrained( "microsoft/Phi-3-mini-4k-instruct" ) pipe = pipeline( "text-generation", model=model, tokenizer=tokenizer ) generation_args = { "max_new_tokens": 500, "return_full_text": False, "temperature": 0.0, "do_sample": False, } # ----------------------- # ルートページ (HTML) # ----------------------- @app.route("/") def index(): return """