Spaces:
Sleeping
Sleeping
| import gradio as gr | |
| import openai | |
| import os | |
| from typing import Optional | |
| from gradio_client import Client | |
| ############################# | |
| # OpenAI API ํด๋ผ์ด์ธํธ ์ค์ | |
| openai.api_key = os.getenv("OPENAI_API_KEY") | |
| if not openai.api_key: | |
| raise ValueError("OpenAI API ํ ํฐ(OPENAI_API_KEY)์ด ์ค์ ๋์ง ์์์ต๋๋ค.") | |
| def call_openai_api( | |
| content: str, | |
| system_message: str, | |
| max_tokens: int, | |
| temperature: float, | |
| top_p: float | |
| ) -> str: | |
| """ | |
| OpenAI์ GPT-4o-mini ๋ชจ๋ธ์ ์ด์ฉํด ํ ๋ฒ์ ์ง๋ฌธ(content)์ ๋ํ ๋ต๋ณ์ ๋ฐํํ๋ ํจ์. | |
| """ | |
| try: | |
| response = openai.ChatCompletion.create( | |
| model="gpt-4o-mini", | |
| messages=[ | |
| {"role": "system", "content": system_message}, | |
| {"role": "user", "content": content}, | |
| ], | |
| max_tokens=max_tokens, | |
| temperature=temperature, | |
| top_p=top_p, | |
| ) | |
| assistant_message = response.choices[0].message['content'] | |
| return assistant_message | |
| except Exception as e: | |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" | |
| ############################# | |
| # ๊ณ ๊ธ ์ค์ (OpenAI) - ์ฝ๋์์๋ง ์ ์ (UI์ ๋ ธ์ถ ๊ธ์ง) | |
| ############################# | |
| OPENAI_SYSTEM_MESSAGE = """๋ฐ๋์ ํ๊ธ๋ก ๋ต๋ณํ ๊ฒ. | |
| ๋๋ ์ต๊ณ ์ ๋น์์ด๋ค. | |
| ๋ด๊ฐ ์๊ตฌํ๋ ๊ฒ๋ค์ ์ต๋ํ ์์ธํ๊ณ ์ ํํ๊ฒ ๋ต๋ณํ๋ผ. | |
| ##[๊ธฐ๋ณธ๊ท์น] | |
| 1. ๋ฐ๋์ ํ๊ตญ์ด(ํ๊ธ)๋ก ์์ฑํ๋ผ. | |
| 2. ๋๋ ๊ฐ์ฅ ์ฃผ๋ชฉ๋ฐ๋ ๋ง์ผํฐ์ด๋ฉฐ ๋ธ๋ก๊ทธ ๋ง์ผํ ์ ๋ฌธ๊ฐ์ด๋ค. | |
| 3. ํนํ ๋๋ '์ ๋ณด์ฑ(Informative)' ์ ๋ฌธ ๋ธ๋ก๊ทธ ๋ง์ผํ ์ ๋ฌธ๊ฐ์ด๋ค. | |
| 4. ์ ๋ณด ์ ๊ณต์ ์ด์ ์ ๋ง์ถ์ด ์์ฑํ๋ค. | |
| ##[ํ ์คํธ ์์ฑ ๊ท์น] | |
| 1. ์์ฃผ์ ๋ฅผ 5๊ฐ๋ก ๊ตฌ๋ถํ์ฌ 2000์ ์ด์๋๋๋ก ์์ฑํ๋ผ. | |
| 2. ์ ์ฒด ๋งฅ๋ฝ์ ์ดํดํ๊ณ ๋ฌธ์ฅ์ ์ผ๊ด์ฑ์ ์ ์งํ๋ผ. | |
| 3. ์ ๋๋ก ์ฐธ๊ณ ๊ธ์ ํ๋ฌธ์ฅ ์ด์ ๊ทธ๋๋ก ์ถ๋ ฅํ์ง ๋ง ๊ฒ. | |
| 4. ์ฃผ์ ์ ์ํฉ์ ๋ง๋ ์ ์ ํ ์ดํ๋ฅผ ์ ํํ๋ผ. | |
| 5. ํ๊ธ ์ดํ์ ๋์ด๋๋ ์ฝ๊ฒ ์์ฑํ๋ผ. | |
| 6. ์ ๋ ๋ฌธ์ฅ์ ๋์ '๋ต๋๋ค'๋ฅผ ์ฌ์ฉํ์ง ๋ง ๊ฒ. | |
| ###[์ ๋ณด์ฑ ๋ธ๋ก๊ทธ ์์ฑ ๊ท์น] | |
| 1. ๋ ์๊ฐ ์ป๊ณ ์ ํ๋ ์ ์ฉํ ์ ๋ณด์ ํฅ๋ฏธ๋ก์ด ์ ๋ณด๋ฅผ ์ ๊ณตํ๋๋ก ์์ฑํ๋ผ. | |
| 2. ๋ ์์ ๊ณต๊ฐ์ ์ด๋์ด๋ด๊ณ ๊ถ๊ธ์ฆ์ ํด๊ฒฐํ๋๋ก ์์ฑํ๋ผ. | |
| 3. ๋ ์์ ๊ด์ฌ์ฌ๋ฅผ ์ถฉ์กฑ์ํค๋๋ก ์์ฑํ๋ผ. | |
| 4. ๋ ์์๊ฒ ์ด๋์ด ๋๋ ์ ๋ณด๋ฅผ ์์ฑํ๋ผ. | |
| ##[์ ์ธ ๊ท์น] | |
| 1. ๋ฐ๋์ ๋น์์ด ๋ฐ ์์ค(expletive, abusive language, slang)์ ์ ์ธํ๋ผ. | |
| 2. ๋ฐ๋์ ์ฐธ๊ณ ๊ธ์ ๋งํฌ(URL)๋ ์ ์ธํ๋ผ. | |
| 3. ์ฐธ๊ณ ๊ธ์์ '๋งํฌ๋ฅผ ํ์ธํด์ฃผ์ธ์'์ ๊ฐ์ ๋งํฌ ์ด๋์ ๋ฌธ๊ตฌ๋ ์ ์ธํ๋ผ. | |
| 4. ์ฐธ๊ณ ๊ธ์ ์๋ ์์ฑ์, ํ์, ์ ํ๋ฒ, ๊ธฐ์์ ์ด๋ฆ, ์ ์นญ, ๋๋ค์์ ๋ฐ๋์ ์ ์ธํ๋ผ. | |
| 5. ๋ฐ๋์ ๋ฌธ์ฅ์ ๋๋ถ๋ถ์ด ์ด์ํ ํ๊ตญ์ด ํํ์ ์ ์ธํ๋ผ('์์', '๋ต๋๋ค', 'ํด์', 'ํด์ฃผ์ฃ ', '๋์ฃ ', '๋์ด์', '๊ณ ์' ๋ฑ.) | |
| """ | |
| OPENAI_MAX_TOKENS = 4000 | |
| OPENAI_TEMPERATURE = 0.7 | |
| OPENAI_TOP_P = 0.95 | |
| ############################# | |
| # API ํด๋ผ์ด์ธํธ ์ค์ (ํ๊น ํ์ด์ค ์คํ์ด์ค) | |
| ############################# | |
| client = Client("Kims12/5-3_N-blog") | |
| ############################# | |
| # UI - ๋ธ๋ก๊ทธ ์์ฑ๊ธฐ | |
| ############################# | |
| with gr.Blocks() as demo: | |
| gr.Markdown("# ๋ธ๋ก๊ทธ ์์ฑ๊ธฐ") | |
| # ๋งํฌ๋ฐ๊พธ๊ธฐ (๋ผ๋์ค ๋ฒํผ) | |
| tone_radio = gr.Radio( | |
| label="๋งํฌ๋ฐ๊พธ๊ธฐ", | |
| choices=["์น๊ทผํ๊ฒ", "์ผ๋ฐ์ ์ธ", "์ ๋ฌธ์ ์ธ"], | |
| value="์ผ๋ฐ์ ์ธ" # ๊ธฐ๋ณธ ์ ํ | |
| ) | |
| # ์ฐธ์กฐ๊ธ ์ ๋ ฅ (3๊ฐ)๊ณผ API ์คํ ๋ฒํผ | |
| with gr.Row(): | |
| with gr.Column(): | |
| ref1 = gr.Textbox(label="์ฐธ์กฐ๊ธ 1") | |
| fetch_button1 = gr.Button("API ์คํ1") | |
| with gr.Column(): | |
| ref2 = gr.Textbox(label="์ฐธ์กฐ๊ธ 2") | |
| fetch_button2 = gr.Button("API ์คํ2") | |
| with gr.Column(): | |
| ref3 = gr.Textbox(label="์ฐธ์กฐ๊ธ 3") | |
| fetch_button3 = gr.Button("API ์คํ3") | |
| output_box = gr.Textbox(label="๊ฒฐ๊ณผ", lines=20, interactive=False) | |
| # ์ฐธ์กฐ๊ธ1 API ์คํ ํจ์ | |
| def fetch_ref1(url: str) -> str: | |
| if not url: | |
| return "URL์ ์ ๋ ฅํด์ฃผ์ธ์." | |
| try: | |
| result = client.predict( | |
| url=url, | |
| api_name="/predict" | |
| ) | |
| return result | |
| except Exception as e: | |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" | |
| # ์ฐธ์กฐ๊ธ2 API ์คํ ํจ์ | |
| def fetch_ref2(url: str) -> str: | |
| if not url: | |
| return "URL์ ์ ๋ ฅํด์ฃผ์ธ์." | |
| try: | |
| result = client.predict( | |
| url=url, | |
| api_name="/predict" | |
| ) | |
| return result | |
| except Exception as e: | |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" | |
| # ์ฐธ์กฐ๊ธ3 API ์คํ ํจ์ | |
| def fetch_ref3(url: str) -> str: | |
| if not url: | |
| return "URL์ ์ ๋ ฅํด์ฃผ์ธ์." | |
| try: | |
| result = client.predict( | |
| url=url, | |
| api_name="/predict" | |
| ) | |
| return result | |
| except Exception as e: | |
| return f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}" | |
| # ๋ฒํผ ํด๋ฆญ ์ ํด๋น ์ฐธ์กฐ๊ธ API ์คํ | |
| fetch_button1.click( | |
| fn=fetch_ref1, | |
| inputs=ref1, | |
| outputs=ref1 | |
| ) | |
| fetch_button2.click( | |
| fn=fetch_ref2, | |
| inputs=ref2, | |
| outputs=ref2 | |
| ) | |
| fetch_button3.click( | |
| fn=fetch_ref3, | |
| inputs=ref3, | |
| outputs=ref3 | |
| ) | |
| # ๋ธ๋ก๊ทธ ์์ฑ ํจ์ | |
| def generate_blog(tone_value: str, ref1_value: str, ref2_value: str, ref3_value: str) -> str: | |
| # ํ๋กฌํํธ ์์ฑ | |
| question = ( | |
| f"๋งํฌ: {tone_value}\n" | |
| f"์ฐธ์กฐ๊ธ1: {ref1_value}\n" | |
| f"์ฐธ์กฐ๊ธ2: {ref2_value}\n" | |
| f"์ฐธ์กฐ๊ธ3: {ref3_value}\n" | |
| ) | |
| # OpenAI GPT-4o-mini ๋ชจ๋ธ ํธ์ถ | |
| response = call_openai_api( | |
| content=question, | |
| system_message=OPENAI_SYSTEM_MESSAGE, | |
| max_tokens=OPENAI_MAX_TOKENS, | |
| temperature=OPENAI_TEMPERATURE, | |
| top_p=OPENAI_TOP_P | |
| ) | |
| return response | |
| generate_button = gr.Button("์์ฑํ๊ธฐ") | |
| generate_button.click( | |
| fn=generate_blog, | |
| inputs=[tone_radio, ref1, ref2, ref3], | |
| outputs=output_box | |
| ) | |
| if __name__ == "__main__": | |
| demo.launch() | |