ADAPT-Chase's picture
Elizabeth data update 2025-08-23 16:57
fd382c9 verified
#!/usr/bin/env python3
"""
HuggingFace Xet Push for Elizabeth Data
Uses HF's built-in Xet integration for efficient data versioning
"""
import os
import json
import shutil
from datetime import datetime
from pathlib import Path
from huggingface_hub import HfApi
import logging
# Set up logging
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)
class ElizabethHFXetPush:
"""Push Elizabeth data to HuggingFace using Xet integration"""
def __init__(self):
self.api = HfApi()
self.dataset_id = "LevelUp2x/elizabeth-data"
self.temp_dir = "/tmp/elizabeth_xet_upload"
# Ensure temp directory exists
os.makedirs(self.temp_dir, exist_ok=True)
def prepare_data(self):
"""Prepare all Elizabeth data for upload"""
logger.info("Preparing Elizabeth data for HF Xet upload...")
# Create data structure
data_structure = {
"version": "1.0",
"timestamp": datetime.now().isoformat(),
"elizabeth_version": "v0.0.2",
"data_sources": []
}
# 1. Database files
db_files = []
if os.path.exists("/workspace/elizabeth_memory.db"):
shutil.copy2("/workspace/elizabeth_memory.db", self.temp_dir)
db_files.append("elizabeth_memory.db")
logger.info("✓ Copied elizabeth_memory.db")
if os.path.exists("/workspace/nova_memory.db"):
shutil.copy2("/workspace/nova_memory.db", self.temp_dir)
db_files.append("nova_memory.db")
logger.info("✓ Copied nova_memory.db")
if db_files:
data_structure["data_sources"].append({
"type": "databases",
"files": db_files,
"description": "SQLite databases with conversation history"
})
# 2. Repository code (lightweight version)
repo_dir = os.path.join(self.temp_dir, "repository")
os.makedirs(repo_dir, exist_ok=True)
# Copy essential repository files
essential_paths = [
"/workspace/elizabeth-repo/versions",
"/workspace/elizabeth-repo/src",
"/workspace/elizabeth-repo/tools",
"/workspace/elizabeth-repo/scripts",
"/workspace/elizabeth-repo/README.md",
"/workspace/elizabeth-repo/requirements.txt"
]
for path in essential_paths:
if os.path.exists(path):
if os.path.isfile(path):
shutil.copy2(path, repo_dir)
else:
dest_path = os.path.join(repo_dir, os.path.basename(path))
shutil.copytree(path, dest_path, dirs_exist_ok=True)
logger.info("✓ Copied repository structure")
data_structure["data_sources"].append({
"type": "code",
"description": "Elizabeth repository with versions v0.0.1 and v0.0.2"
})
# 3. Create manifest
manifest_path = os.path.join(self.temp_dir, "manifest.json")
with open(manifest_path, 'w') as f:
json.dump(data_structure, f, indent=2)
logger.info("✓ Created data manifest")
return data_structure
def upload_to_hf(self, commit_message=None):
"""Upload prepared data to HuggingFace dataset"""
if not commit_message:
commit_message = f"Elizabeth data update {datetime.now().strftime('%Y-%m-%d %H:%M')}"
try:
logger.info(f"Uploading to HuggingFace dataset: {self.dataset_id}")
# Check authentication
try:
self.api.whoami()
except Exception as auth_error:
logger.error(f"Authentication failed: {auth_error}")
logger.error("Please set HF_TOKEN environment variable:")
logger.error("export HF_TOKEN='your_huggingface_token_here'")
logger.error("Or login with: huggingface-cli login")
return {
"success": False,
"error": f"Authentication required: {auth_error}",
"instructions": "Set HF_TOKEN environment variable or run 'huggingface-cli login'"
}
# Create dataset if it doesn't exist
try:
self.api.dataset_info(self.dataset_id)
except:
logger.info("Dataset doesn't exist, creating...")
self.api.create_repo(
self.dataset_id,
repo_type="dataset"
)
# Upload files
self.api.upload_folder(
folder_path=self.temp_dir,
repo_id=self.dataset_id,
repo_type="dataset",
commit_message=commit_message,
# HF Xet will automatically handle efficient uploads
)
logger.info("✅ Upload completed successfully!")
logger.info(f"Dataset URL: https://huggingface.co/datasets/{self.dataset_id}")
return {
"success": True,
"dataset_url": f"https://huggingface.co/datasets/{self.dataset_id}",
"commit_message": commit_message
}
except Exception as e:
logger.error(f"Upload failed: {e}")
return {
"success": False,
"error": str(e)
}
def cleanup(self):
"""Clean up temporary files"""
if os.path.exists(self.temp_dir):
shutil.rmtree(self.temp_dir)
logger.info("Cleaned up temporary files")
def run_full_upload(self):
"""Run complete upload process"""
try:
# Prepare data
data_info = self.prepare_data()
# Upload to HF
result = self.upload_to_hf()
# Cleanup
self.cleanup()
return {
"preparation": data_info,
"upload": result
}
except Exception as e:
self.cleanup()
return {
"success": False,
"error": str(e)
}
def main():
"""Command line interface"""
import argparse
parser = argparse.ArgumentParser(description="Elizabeth HF Xet Upload")
parser.add_argument("--upload", action="store_true", help="Upload data to HuggingFace")
parser.add_argument("--prepare-only", action="store_true", help="Only prepare data, don't upload")
parser.add_argument("--commit-message", help="Custom commit message")
args = parser.parse_args()
uploader = ElizabethHFXetPush()
if args.prepare_only:
# Just prepare data
data_info = uploader.prepare_data()
print("Data prepared at:", uploader.temp_dir)
print("Manifest:")
print(json.dumps(data_info, indent=2))
elif args.upload:
# Full upload
result = uploader.run_full_upload()
print("Upload result:")
print(json.dumps(result, indent=2))
else:
# Show info
print("Elizabeth HF Xet Upload Tool")
print("Dataset:", uploader.dataset_id)
print("Usage: python hf_xet_push.py --upload")
print("Options: --prepare-only, --commit-message 'Custom message'")
if __name__ == "__main__":
main()