Restaurantguide / s3bucket_pathadder.py
dhanvanth183's picture
Upload 16 files
267bd16 verified
import pandas as pd
import os
# Load the CSV file
file_path = "D:/Projects/Liminal/AI_Guide/resources/Dataset.csv" # Update with your input CSV file path
df = pd.read_csv(file_path)
# Define S3 bucket details
bucket_name = "restaurant-resources-liminal"
region = "ap-south-1"
s3_base_url = f"https://{bucket_name}.s3.{region}.amazonaws.com/uuid_images/"
# Update the Image_path column with S3 URLs using the Unique_ID column
df['S3Image_path'] = df['Unique_ID'].apply(lambda x: f"{s3_base_url}{x}.png")
# Save the updated DataFrame back to a new CSV file
#updated_csv_path = "updated_restaurant_data.csv"
#df.to_csv(updated_csv_path, index=False)
#print(f"Updated CSV file saved to: {updated_csv_path}")
"""
# Function to extract the file type and build S3 URL
def generate_s3_url(image_path, unique_id):
# Extract file extension dynamically
file_extension = os.path.splitext(image_path)[-1] # E.g., '.jpg', '.png'
# Build the full S3 URL
return f"{s3_base_url}{unique_id}{file_extension}"""
# Apply the function to create a new 'S3_Image_URL' column
#df['S3_Image_URL'] = df.apply(lambda row: generate_s3_url(row['Image_path'], row['Unique_ID']), axis=1)
# Save the updated DataFrame to a new CSV
output_file_path = 'D:/Projects/Liminal/AI_Guide/resources/S3Dataset.csv'
df.to_csv(output_file_path, index=False)
print(f"Updated CSV file saved at: {output_file_path}")