File size: 1,431 Bytes
267bd16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
import pandas as pd
import os

# Load the CSV file
file_path = "D:/Projects/Liminal/AI_Guide/resources/Dataset.csv"  # Update with your input CSV file path
df = pd.read_csv(file_path)

# Define S3 bucket details
bucket_name = "restaurant-resources-liminal"
region = "ap-south-1"
s3_base_url = f"https://{bucket_name}.s3.{region}.amazonaws.com/uuid_images/"

# Update the Image_path column with S3 URLs using the Unique_ID column
df['S3Image_path'] = df['Unique_ID'].apply(lambda x: f"{s3_base_url}{x}.png")

# Save the updated DataFrame back to a new CSV file
#updated_csv_path = "updated_restaurant_data.csv"
#df.to_csv(updated_csv_path, index=False)

#print(f"Updated CSV file saved to: {updated_csv_path}")
"""

# Function to extract the file type and build S3 URL

def generate_s3_url(image_path, unique_id):

    # Extract file extension dynamically

    file_extension = os.path.splitext(image_path)[-1]  # E.g., '.jpg', '.png'

    # Build the full S3 URL

    return f"{s3_base_url}{unique_id}{file_extension}"""

# Apply the function to create a new 'S3_Image_URL' column
#df['S3_Image_URL'] = df.apply(lambda row: generate_s3_url(row['Image_path'], row['Unique_ID']), axis=1)

# Save the updated DataFrame to a new CSV
output_file_path = 'D:/Projects/Liminal/AI_Guide/resources/S3Dataset.csv'
df.to_csv(output_file_path, index=False)

print(f"Updated CSV file saved at: {output_file_path}")