Update llm_inference.py
Browse files- llm_inference.py +1 -17
llm_inference.py
CHANGED
|
@@ -22,7 +22,6 @@ class LLMInferenceNode:
|
|
| 22 |
api_key=self.sambanova_api_key,
|
| 23 |
base_url="https://api.sambanova.ai/v1",
|
| 24 |
)
|
| 25 |
-
self.flux_client = Client("KingNish/Realtime-FLUX", hf_token=self.huggingface_token)
|
| 26 |
|
| 27 |
def generate_prompt(self, dynamic_seed, prompt_type, custom_input):
|
| 28 |
"""
|
|
@@ -242,19 +241,4 @@ Your output is only the caption itself, no comments or extra formatting. The cap
|
|
| 242 |
|
| 243 |
except Exception as e:
|
| 244 |
print(f"An error occurred: {e}")
|
| 245 |
-
return f"Error occurred while processing the request: {str(e)}"
|
| 246 |
-
|
| 247 |
-
def generate_image(self, prompt, seed=42, width=1024, height=1024):
|
| 248 |
-
try:
|
| 249 |
-
result = self.flux_client.predict(
|
| 250 |
-
prompt=prompt,
|
| 251 |
-
seed=seed,
|
| 252 |
-
width=width,
|
| 253 |
-
height=height,
|
| 254 |
-
api_name="/generate_image"
|
| 255 |
-
)
|
| 256 |
-
# Extract the image path from the result tuple
|
| 257 |
-
image_path = result[0]
|
| 258 |
-
return image_path
|
| 259 |
-
except Exception as e:
|
| 260 |
-
raise Exception(f"Error generating image: {str(e)}")
|
|
|
|
| 22 |
api_key=self.sambanova_api_key,
|
| 23 |
base_url="https://api.sambanova.ai/v1",
|
| 24 |
)
|
|
|
|
| 25 |
|
| 26 |
def generate_prompt(self, dynamic_seed, prompt_type, custom_input):
|
| 27 |
"""
|
|
|
|
| 241 |
|
| 242 |
except Exception as e:
|
| 243 |
print(f"An error occurred: {e}")
|
| 244 |
+
return f"Error occurred while processing the request: {str(e)}"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|