# NCTC import os import shutil import gradio as gr from transformers import ReactCodeAgent, HfEngine, Tool import pandas as pd from gradio import Chatbot from transformers.agents import stream_to_gradio from huggingface_hub import login from gradio.data_classes import FileData login(os.getenv("HUGGINGFACEHUB_API_TOKEN")) llm_engine = HfEngine("meta-llama/Meta-Llama-3.1-70B-Instruct") agent = ReactCodeAgent( tools=[], llm_engine=llm_engine, additional_authorized_imports=["numpy", "pandas", "matplotlib.pyplot", "seaborn", "scipy.stats"], max_iterations=10, ) # base_prompt = """You are an expert data analyst of National Customs Targeting Center. You will be uploaded with CSV file with multiple columns of numerical , categorical and text variables. # According to the features you have and the data structure given below, determine which feature should be the target. # Then list 3 interesting questions that could be asked on this data, for instance about specific correlations with target variable. # Then answer these questions one by one, by finding the relevant numbers. # Meanwhile, plot some figures using matplotlib/seaborn and save them to the (already existing) folder './figures/': take care to clear each figure with plt.clf() before doing another plot. # In your final answer: summarize these correlations and trends # After each number derive real worlds insights, for instance: "Correlation between is_december and boredness is 1.3453, which suggest people are more bored in winter". # Your final answer should be a long string with at least 3 numbered and detailed parts. # Structure of the data: # {structure_notes} # The data file is passed to you as the variable data_file, it is a pandas dataframe, you can use it directly. # DO NOT try to load data_file, it is already a dataframe pre-loaded in your python interpreter! # """ base_prompt = """You are an expert data analyst at the National Customs Targeting Center. You will be provided with a CSV file containing multiple columns of numerical, categorical, and text variables. Your tasks are: 1. **Target Identification**: - Determine which feature(s) should be the target for analysis. Focus primarily on numerical and categorical columns, and avoid using unstructured text columns as targets. 2. **Generate Interesting Questions**: - Based on the identified target features, list at least 3 interesting questions that could be asked. For instance, explore specific correlations with the target variable(s), trends, or patterns. 3. **Answer the Questions**: - Answer these questions one by one by analyzing the data and finding relevant numbers. - Generate insights from these answers. For example: "Correlation between `is_december` and `boredness` is 1.3453, suggesting that people are more bored in winter." 4. **Generate Outlier Insights**: - Identify outliers for each variable in the dataset. - Provide insights into the outliers, including printing the outlier records and explaining their significance. 5. **Visualization**: - Plot multiple figures using matplotlib or seaborn. - Generate plots for various target columns, covering both numerical and categorical columns. - Ensure each figure is saved to the './figures/' folder and clear each figure with `plt.clf()` before generating the next plot. - Include relevant plots that visualize correlations, trends, distributions, and outliers. 6. **Final Summary**: - Summarize the correlations, trends, and outlier insights in a detailed manner. Provide at least 3 numbered and detailed parts in the summary. Structure of the data: {structure_notes} The data file is passed to you as the variable `data_file`, which is a pandas dataframe, and you can use it directly. DO NOT try to load `data_file`, as it is already pre-loaded in your Python interpreter! Your final output should include: 1. The identified target feature(s). 2. Three interesting questions with detailed answers and real-world insights. 3. Outlier insights for each variable, including the outlier records. 4. Multiple saved plots in the './figures/' folder. 5. A long, detailed final summary. """ example_notes="""This data is about a sample Customs dataset with products imports (IMP_DESC),Importer ID( IEC No.), SUPPLIER ID , (item unit price) ITEM_UPI , CTH for product classification (Declared CTH), declared BCD Notification benefit (BCD Notification No. Declared) and value of import (ITEM_ASSESS_VAL)""" def get_images_in_directory(directory): image_extensions = {'.png', '.jpg', '.jpeg', '.gif', '.bmp', '.tiff'} image_files = [] for root, dirs, files in os.walk(directory): for file in files: if os.path.splitext(file)[1].lower() in image_extensions: image_files.append(os.path.join(root, file)) return image_files def interact_with_agent(file_input, additional_notes): shutil.rmtree("./figures") os.makedirs("./figures") data_file = pd.read_csv(file_input) data_structure_notes = f"""- Description (output of .describe()): {data_file.describe()} - Columns with dtypes: {data_file.dtypes}""" prompt = base_prompt.format(structure_notes=data_structure_notes) if additional_notes and len(additional_notes) > 0: prompt += "\nAdditional notes on the data:\n" + additional_notes messages = [gr.ChatMessage(role="user", content=prompt)] yield messages + [ gr.ChatMessage(role="assistant", content="⏳ _Starting task..._") ] plot_image_paths = {} for msg in stream_to_gradio(agent, prompt, data_file=data_file): messages.append(msg) for image_path in get_images_in_directory("./figures"): if image_path not in plot_image_paths: image_message = gr.ChatMessage( role="assistant", content=FileData(path=image_path, mime_type="image/png"), ) plot_image_paths[image_path] = True messages.append(image_message) yield messages + [ gr.ChatMessage(role="assistant", content="⏳ _Still processing..._") ] yield messages import gradio as gr with gr.Blocks( theme=gr.themes.Soft( primary_hue=gr.themes.colors.green, # Changing to a fresh green secondary_hue=gr.themes.colors.purple, # Adding a touch of regal purple ) ) as demo: gr.Markdown("""

NCTC Llama-3.1 Data Analysis Agent 📊🤔

NCTC's attempt to use LLM-based ReAct Autonomous Agents to assist in smart customs data analysis

Drop a .csv file below, add notes to describe this data if needed, and Llama-3.1-70B will analyze the file content and draw figures for you!

""") file_input = gr.File(label="Your file to analyze") text_input = gr.Textbox( label="Additional notes to support the analysis" ) submit = gr.Button("Run analysis!", variant="primary") chatbot = gr.Chatbot( label="Data Analyst Agent", type="messages", avatar_images=( None, "https://em-content.zobj.net/source/twitter/53/robot-face_1f916.png", ), ) gr.Examples( examples=[["./example/sample_customs_data_anonymised.csv", example_notes]], inputs=[file_input, text_input], cache_examples=False ) submit.click(interact_with_agent, [file_input, text_input], [chatbot]) if __name__ == "__main__": demo.launch()