import os import streamlit as st from pathlib import Path from datasets import load_dataset from dotenv import load_dotenv def list_files_in_directory(directory): """List all files in the given directory and its subdirectories.""" files = [] for root, dirs, filenames in os.walk(directory): for filename in filenames: files.append(os.path.join(root, filename)) return files def main(): st.title("Embeddings File Browser") # Load environment variables load_dotenv() # Retrieve the Hugging Face token hf_token = os.getenv("HF_TOKEN") if not hf_token: st.error("HF_TOKEN not found in environment variables.") return # Load the dataset using the token try: dataset = load_dataset("SongLift/LyrGen2_DB", use_auth_token=hf_token) st.write("Dataset loaded successfully.") except Exception as e: st.error(f"Error loading dataset: {str(e)}") return # Directory to browse directory = "/data" # Persistent storage directory st.write(f"Browsing directory: {directory}") # List files files = list_files_in_directory(directory) if files: st.write("Files found:") for file in files: st.write(file) else: st.write("No files found in the directory.") if __name__ == "__main__": main()