Spaces:
Running
Running
Commit
·
ed60c89
1
Parent(s):
e6f9ae7
trying fix tenant issue
Browse files- langchain_helper.py +6 -5
langchain_helper.py
CHANGED
|
@@ -33,16 +33,16 @@ current_model_id = os.getenv('model_id')
|
|
| 33 |
def get_few_shot_db_chain(user_message):
|
| 34 |
#llm = AzureOpenAI(deployment_name="gpt-35-turbo-instruct", temperature=0.2)
|
| 35 |
llm = ChatOpenAI(model = current_model_id)
|
| 36 |
-
|
| 37 |
engine = create_engine("sqlite:///ecomm.db")
|
| 38 |
db = SQLDatabase(engine=engine, sample_rows_in_table_info=3)
|
| 39 |
|
| 40 |
embeddings = OpenAIEmbeddings(model="text-embedding-3-small")
|
| 41 |
-
|
| 42 |
to_vectorize = [" ".join(example.values()) for example in few_shots]
|
| 43 |
|
| 44 |
vectorstore = Chroma.from_texts(to_vectorize, embeddings, metadatas=few_shots)
|
| 45 |
-
|
| 46 |
example_selector = SemanticSimilarityExampleSelector(vectorstore=vectorstore, k=2)
|
| 47 |
|
| 48 |
example_prompt = PromptTemplate(
|
|
@@ -59,7 +59,7 @@ def get_few_shot_db_chain(user_message):
|
|
| 59 |
)
|
| 60 |
|
| 61 |
chain = SQLDatabaseChain.from_llm(llm, db, verbose=True, prompt=few_shot_prompt, return_intermediate_steps = True)
|
| 62 |
-
|
| 63 |
response_llm = chain.invoke(user_message)
|
| 64 |
|
| 65 |
print(f"sql query : {response_llm['intermediate_steps'][1]}")
|
|
@@ -67,7 +67,8 @@ def get_few_shot_db_chain(user_message):
|
|
| 67 |
intermediate_sql_query = response_llm['intermediate_steps'][2]['sql_cmd']
|
| 68 |
|
| 69 |
result_df = pd.read_sql_query(intermediate_sql_query, engine)
|
| 70 |
-
|
|
|
|
| 71 |
output_dict = {
|
| 72 |
"result_df" : result_df,
|
| 73 |
"sql_command" : intermediate_sql_query,
|
|
|
|
| 33 |
def get_few_shot_db_chain(user_message):
|
| 34 |
#llm = AzureOpenAI(deployment_name="gpt-35-turbo-instruct", temperature=0.2)
|
| 35 |
llm = ChatOpenAI(model = current_model_id)
|
| 36 |
+
print(llm)
|
| 37 |
engine = create_engine("sqlite:///ecomm.db")
|
| 38 |
db = SQLDatabase(engine=engine, sample_rows_in_table_info=3)
|
| 39 |
|
| 40 |
embeddings = OpenAIEmbeddings(model="text-embedding-3-small")
|
| 41 |
+
print(embeddings)
|
| 42 |
to_vectorize = [" ".join(example.values()) for example in few_shots]
|
| 43 |
|
| 44 |
vectorstore = Chroma.from_texts(to_vectorize, embeddings, metadatas=few_shots)
|
| 45 |
+
print(vectorstore)
|
| 46 |
example_selector = SemanticSimilarityExampleSelector(vectorstore=vectorstore, k=2)
|
| 47 |
|
| 48 |
example_prompt = PromptTemplate(
|
|
|
|
| 59 |
)
|
| 60 |
|
| 61 |
chain = SQLDatabaseChain.from_llm(llm, db, verbose=True, prompt=few_shot_prompt, return_intermediate_steps = True)
|
| 62 |
+
print(chain)
|
| 63 |
response_llm = chain.invoke(user_message)
|
| 64 |
|
| 65 |
print(f"sql query : {response_llm['intermediate_steps'][1]}")
|
|
|
|
| 67 |
intermediate_sql_query = response_llm['intermediate_steps'][2]['sql_cmd']
|
| 68 |
|
| 69 |
result_df = pd.read_sql_query(intermediate_sql_query, engine)
|
| 70 |
+
print("Printing results")
|
| 71 |
+
print(result_df)
|
| 72 |
output_dict = {
|
| 73 |
"result_df" : result_df,
|
| 74 |
"sql_command" : intermediate_sql_query,
|