Prathamesh Sarjerao Vaidya
commited on
Commit
·
65fbbac
1
Parent(s):
65f46e8
made changes
Browse files- .github/workflows/check.yml +58 -3
- .github/workflows/main.yml +48 -3
- Dockerfile +6 -0
- model_preloader.py +57 -27
- startup.py +30 -8
.github/workflows/check.yml
CHANGED
|
@@ -20,28 +20,77 @@ jobs:
|
|
| 20 |
runs-on: ubuntu-latest
|
| 21 |
needs: check-file-size
|
| 22 |
if: github.event_name == 'pull_request'
|
|
|
|
|
|
|
| 23 |
steps:
|
| 24 |
- uses: actions/checkout@v3
|
| 25 |
with:
|
| 26 |
lfs: true
|
|
|
|
| 27 |
|
| 28 |
- name: Pull LFS files
|
| 29 |
run: |
|
| 30 |
git lfs install
|
| 31 |
git lfs pull
|
| 32 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
- name: Setup Python
|
| 34 |
uses: actions/setup-python@v4
|
| 35 |
with:
|
| 36 |
python-version: '3.11'
|
| 37 |
|
|
|
|
| 38 |
- name: Setup system dependencies
|
| 39 |
-
|
|
|
|
|
|
|
|
|
|
| 40 |
|
|
|
|
| 41 |
- name: Convert MD to PDF
|
| 42 |
-
|
|
|
|
|
|
|
|
|
|
| 43 |
|
|
|
|
| 44 |
- name: Upload PDF artifacts
|
|
|
|
| 45 |
uses: actions/upload-artifact@v4
|
| 46 |
with:
|
| 47 |
name: converted-pdfs
|
|
@@ -51,4 +100,10 @@ jobs:
|
|
| 51 |
- name: Upload to Google Drive
|
| 52 |
env:
|
| 53 |
GOOGLE_OAUTH_TOKEN: ${{ secrets.GOOGLE_OAUTH_TOKEN }}
|
| 54 |
-
run:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
runs-on: ubuntu-latest
|
| 21 |
needs: check-file-size
|
| 22 |
if: github.event_name == 'pull_request'
|
| 23 |
+
outputs:
|
| 24 |
+
skip_pdf: ${{ steps.check_md_changes.outputs.skip_pdf }}
|
| 25 |
steps:
|
| 26 |
- uses: actions/checkout@v3
|
| 27 |
with:
|
| 28 |
lfs: true
|
| 29 |
+
fetch-depth: 0 # Need full history for git diff
|
| 30 |
|
| 31 |
- name: Pull LFS files
|
| 32 |
run: |
|
| 33 |
git lfs install
|
| 34 |
git lfs pull
|
| 35 |
|
| 36 |
+
# NEW STEP: Check if MD files were modified
|
| 37 |
+
- name: Check for MD file changes
|
| 38 |
+
id: check_md_changes
|
| 39 |
+
run: |
|
| 40 |
+
echo "Checking for markdown file changes..."
|
| 41 |
+
|
| 42 |
+
# Get the commit message
|
| 43 |
+
COMMIT_MSG="${{ github.event.head_commit.message }}"
|
| 44 |
+
if [ -z "$COMMIT_MSG" ]; then
|
| 45 |
+
COMMIT_MSG=$(git log -1 --pretty=%B)
|
| 46 |
+
fi
|
| 47 |
+
echo "Commit message: $COMMIT_MSG"
|
| 48 |
+
|
| 49 |
+
# Check if commit message indicates MD changes
|
| 50 |
+
MD_IN_COMMIT=$(echo "$COMMIT_MSG" | grep -i "\.md\|markdown\|documentation\|docs\|readme" || true)
|
| 51 |
+
|
| 52 |
+
# Check if any MD files were actually modified in the diff
|
| 53 |
+
if [ "${{ github.event_name }}" = "pull_request" ]; then
|
| 54 |
+
MD_FILES_CHANGED=$(git diff --name-only origin/main...HEAD | grep "\.md$" || true)
|
| 55 |
+
else
|
| 56 |
+
MD_FILES_CHANGED=$(git diff --name-only HEAD~1 HEAD | grep "\.md$" || true)
|
| 57 |
+
fi
|
| 58 |
+
|
| 59 |
+
echo "MD files in commit message: $MD_IN_COMMIT"
|
| 60 |
+
echo "MD files changed: $MD_FILES_CHANGED"
|
| 61 |
+
|
| 62 |
+
# Skip PDF conversion if no MD files changed AND no MD-related keywords in commit
|
| 63 |
+
if [ -z "$MD_FILES_CHANGED" ] && [ -z "$MD_IN_COMMIT" ]; then
|
| 64 |
+
echo "skip_pdf=true" >> $GITHUB_OUTPUT
|
| 65 |
+
echo "Skipping PDF conversion - no MD files modified"
|
| 66 |
+
else
|
| 67 |
+
echo "skip_pdf=false" >> $GITHUB_OUTPUT
|
| 68 |
+
echo "MD files detected - will convert to PDF"
|
| 69 |
+
echo "Changed MD files: $MD_FILES_CHANGED"
|
| 70 |
+
fi
|
| 71 |
+
|
| 72 |
- name: Setup Python
|
| 73 |
uses: actions/setup-python@v4
|
| 74 |
with:
|
| 75 |
python-version: '3.11'
|
| 76 |
|
| 77 |
+
# CONDITIONAL STEP: Only run if MD files changed
|
| 78 |
- name: Setup system dependencies
|
| 79 |
+
if: steps.check_md_changes.outputs.skip_pdf == 'false'
|
| 80 |
+
run: |
|
| 81 |
+
echo "Setting up system dependencies for PDF conversion..."
|
| 82 |
+
chmod +x .github/workflows/scripts/setup_system.sh && .github/workflows/scripts/setup_system.sh
|
| 83 |
|
| 84 |
+
# CONDITIONAL STEP: Only run if MD files changed
|
| 85 |
- name: Convert MD to PDF
|
| 86 |
+
if: steps.check_md_changes.outputs.skip_pdf == 'false'
|
| 87 |
+
run: |
|
| 88 |
+
echo "Converting MD files to PDF..."
|
| 89 |
+
chmod +x .github/workflows/scripts/convert_md_to_pdf.sh && .github/workflows/scripts/convert_md_to_pdf.sh
|
| 90 |
|
| 91 |
+
# CONDITIONAL STEP: Only run if MD files changed
|
| 92 |
- name: Upload PDF artifacts
|
| 93 |
+
if: steps.check_md_changes.outputs.skip_pdf == 'false'
|
| 94 |
uses: actions/upload-artifact@v4
|
| 95 |
with:
|
| 96 |
name: converted-pdfs
|
|
|
|
| 100 |
- name: Upload to Google Drive
|
| 101 |
env:
|
| 102 |
GOOGLE_OAUTH_TOKEN: ${{ secrets.GOOGLE_OAUTH_TOKEN }}
|
| 103 |
+
run: |
|
| 104 |
+
if [ "${{ steps.check_md_changes.outputs.skip_pdf }}" = "true" ]; then
|
| 105 |
+
echo "Skipped PDF conversion - uploading existing files only"
|
| 106 |
+
else
|
| 107 |
+
echo "Uploading files including new PDFs to Google Drive"
|
| 108 |
+
fi
|
| 109 |
+
python .github/workflows/scripts/upload_to_drive.py
|
.github/workflows/main.yml
CHANGED
|
@@ -18,18 +18,57 @@ jobs:
|
|
| 18 |
git lfs install
|
| 19 |
git lfs pull
|
| 20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
- name: Setup Python
|
| 22 |
uses: actions/setup-python@v4
|
| 23 |
with:
|
| 24 |
python-version: '3.11'
|
| 25 |
|
|
|
|
| 26 |
- name: Setup system dependencies
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
| 28 |
|
|
|
|
| 29 |
- name: Convert MD to PDF
|
| 30 |
-
|
|
|
|
|
|
|
|
|
|
| 31 |
|
|
|
|
| 32 |
- name: Upload PDF artifacts
|
|
|
|
| 33 |
uses: actions/upload-artifact@v4
|
| 34 |
with:
|
| 35 |
name: converted-pdfs
|
|
@@ -42,7 +81,13 @@ jobs:
|
|
| 42 |
GOOGLE_OAUTH_TOKEN: ${{ secrets.GOOGLE_OAUTH_TOKEN }}
|
| 43 |
# Fallback authentication method (Service Account)
|
| 44 |
GOOGLE_SERVICE_ACCOUNT_KEY: ${{ secrets.GOOGLE_SERVICE_ACCOUNT_KEY }}
|
| 45 |
-
run:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
|
| 47 |
- name: Push to Hugging Face hub
|
| 48 |
env:
|
|
|
|
| 18 |
git lfs install
|
| 19 |
git lfs pull
|
| 20 |
|
| 21 |
+
# NEW STEP: Check for MD file changes
|
| 22 |
+
- name: Check for MD file changes
|
| 23 |
+
id: check_md_changes
|
| 24 |
+
run: |
|
| 25 |
+
echo "Checking for markdown file changes..."
|
| 26 |
+
|
| 27 |
+
# Get the commit message
|
| 28 |
+
COMMIT_MSG=$(git log -1 --pretty=%B)
|
| 29 |
+
echo "Commit message: $COMMIT_MSG"
|
| 30 |
+
|
| 31 |
+
# Check if commit message indicates MD changes
|
| 32 |
+
MD_IN_COMMIT=$(echo "$COMMIT_MSG" | grep -i "\.md\|markdown\|documentation\|docs\|readme" || true)
|
| 33 |
+
|
| 34 |
+
# Check if any MD files were actually modified in the last commit
|
| 35 |
+
MD_FILES_CHANGED=$(git diff --name-only HEAD~1 HEAD | grep "\.md$" || true)
|
| 36 |
+
|
| 37 |
+
echo "MD files in commit message: $MD_IN_COMMIT"
|
| 38 |
+
echo "MD files changed: $MD_FILES_CHANGED"
|
| 39 |
+
|
| 40 |
+
# Skip PDF conversion if no MD files changed AND no MD-related keywords in commit
|
| 41 |
+
if [ -z "$MD_FILES_CHANGED" ] && [ -z "$MD_IN_COMMIT" ]; then
|
| 42 |
+
echo "skip_pdf=true" >> $GITHUB_OUTPUT
|
| 43 |
+
echo "Skipping PDF conversion - no MD files modified"
|
| 44 |
+
else
|
| 45 |
+
echo "skip_pdf=false" >> $GITHUB_OUTPUT
|
| 46 |
+
echo "MD files detected - will convert to PDF"
|
| 47 |
+
echo "Changed MD files: $MD_FILES_CHANGED"
|
| 48 |
+
fi
|
| 49 |
+
|
| 50 |
- name: Setup Python
|
| 51 |
uses: actions/setup-python@v4
|
| 52 |
with:
|
| 53 |
python-version: '3.11'
|
| 54 |
|
| 55 |
+
# CONDITIONAL STEP: Only run if MD files changed
|
| 56 |
- name: Setup system dependencies
|
| 57 |
+
if: steps.check_md_changes.outputs.skip_pdf == 'false'
|
| 58 |
+
run: |
|
| 59 |
+
echo "Setting up system dependencies for PDF conversion..."
|
| 60 |
+
chmod +x .github/workflows/scripts/setup_system.sh && .github/workflows/scripts/setup_system.sh
|
| 61 |
|
| 62 |
+
# CONDITIONAL STEP: Only run if MD files changed
|
| 63 |
- name: Convert MD to PDF
|
| 64 |
+
if: steps.check_md_changes.outputs.skip_pdf == 'false'
|
| 65 |
+
run: |
|
| 66 |
+
echo "Converting MD files to PDF..."
|
| 67 |
+
chmod +x .github/workflows/scripts/convert_md_to_pdf.sh && .github/workflows/scripts/convert_md_to_pdf.sh
|
| 68 |
|
| 69 |
+
# CONDITIONAL STEP: Only run if MD files changed
|
| 70 |
- name: Upload PDF artifacts
|
| 71 |
+
if: steps.check_md_changes.outputs.skip_pdf == 'false'
|
| 72 |
uses: actions/upload-artifact@v4
|
| 73 |
with:
|
| 74 |
name: converted-pdfs
|
|
|
|
| 81 |
GOOGLE_OAUTH_TOKEN: ${{ secrets.GOOGLE_OAUTH_TOKEN }}
|
| 82 |
# Fallback authentication method (Service Account)
|
| 83 |
GOOGLE_SERVICE_ACCOUNT_KEY: ${{ secrets.GOOGLE_SERVICE_ACCOUNT_KEY }}
|
| 84 |
+
run: |
|
| 85 |
+
if [ "${{ steps.check_md_changes.outputs.skip_pdf }}" = "true" ]; then
|
| 86 |
+
echo "Skipped PDF conversion - uploading existing files only"
|
| 87 |
+
else
|
| 88 |
+
echo "Uploading files including new PDFs to Google Drive"
|
| 89 |
+
fi
|
| 90 |
+
python .github/workflows/scripts/upload_to_drive.py
|
| 91 |
|
| 92 |
- name: Push to Hugging Face hub
|
| 93 |
env:
|
Dockerfile
CHANGED
|
@@ -26,6 +26,7 @@ RUN apt-get update && apt-get install -y \
|
|
| 26 |
libavformat-dev \
|
| 27 |
libavutil-dev \
|
| 28 |
libswresample-dev \
|
|
|
|
| 29 |
&& rm -rf /var/lib/apt/lists/*
|
| 30 |
|
| 31 |
# Copy requirements first for better caching
|
|
@@ -33,6 +34,11 @@ COPY requirements.txt .
|
|
| 33 |
|
| 34 |
# Install Python dependencies with proper error handling
|
| 35 |
RUN pip install --no-cache-dir --upgrade pip setuptools wheel && \
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
pip install --no-cache-dir -r requirements.txt
|
| 37 |
|
| 38 |
# Copy application code
|
|
|
|
| 26 |
libavformat-dev \
|
| 27 |
libavutil-dev \
|
| 28 |
libswresample-dev \
|
| 29 |
+
execstack \
|
| 30 |
&& rm -rf /var/lib/apt/lists/*
|
| 31 |
|
| 32 |
# Copy requirements first for better caching
|
|
|
|
| 34 |
|
| 35 |
# Install Python dependencies with proper error handling
|
| 36 |
RUN pip install --no-cache-dir --upgrade pip setuptools wheel && \
|
| 37 |
+
# Install ONNX Runtime CPU version specifically
|
| 38 |
+
pip install --no-cache-dir onnxruntime==1.16.3 && \
|
| 39 |
+
# Fix executable stack issue
|
| 40 |
+
find /usr/local/lib/python*/site-packages/onnxruntime -name "*.so" -exec execstack -c {} \; 2>/dev/null || true && \
|
| 41 |
+
# Install other requirements
|
| 42 |
pip install --no-cache-dir -r requirements.txt
|
| 43 |
|
| 44 |
# Copy application code
|
model_preloader.py
CHANGED
|
@@ -397,55 +397,85 @@ class ModelPreloader:
|
|
| 397 |
except Exception as e:
|
| 398 |
logger.warning(f"Error saving cache for {model_key}: {e}")
|
| 399 |
|
| 400 |
-
def load_pyannote_pipeline(self
|
| 401 |
"""Load pyannote speaker diarization pipeline with container-safe settings."""
|
| 402 |
try:
|
| 403 |
console.print(f"[yellow]Loading pyannote.audio pipeline...[/yellow]")
|
| 404 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 405 |
# Check for HuggingFace token
|
| 406 |
-
hf_token = os.getenv('HUGGINGFACE_TOKEN')
|
| 407 |
if not hf_token:
|
| 408 |
console.print("[red]Warning: HUGGINGFACE_TOKEN not found. Some models may not be accessible.[/red]")
|
| 409 |
|
| 410 |
-
#
|
| 411 |
-
|
| 412 |
-
|
| 413 |
-
|
| 414 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 415 |
pipeline = Pipeline.from_pretrained(
|
| 416 |
"pyannote/speaker-diarization-3.1",
|
| 417 |
use_auth_token=hf_token,
|
| 418 |
cache_dir=str(self.cache_dir / "pyannote")
|
| 419 |
)
|
| 420 |
|
| 421 |
-
# Force CPU execution
|
| 422 |
if hasattr(pipeline, '_models'):
|
| 423 |
for model_name, model in pipeline._models.items():
|
| 424 |
if hasattr(model, 'to'):
|
| 425 |
model.to('cpu')
|
| 426 |
-
|
| 427 |
-
|
| 428 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 429 |
|
| 430 |
except Exception as e:
|
| 431 |
-
|
| 432 |
-
if "executable stack" in
|
| 433 |
-
console.print("[yellow]ONNX Runtime
|
| 434 |
-
|
|
|
|
| 435 |
try:
|
| 436 |
-
|
| 437 |
-
|
| 438 |
-
|
| 439 |
-
|
| 440 |
-
|
| 441 |
-
|
| 442 |
-
|
| 443 |
-
)
|
| 444 |
-
|
| 445 |
-
|
| 446 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 447 |
|
| 448 |
-
console.print(f"[red]ERROR: Failed to load pyannote.audio pipeline: {e}[/red]")
|
| 449 |
logger.error(f"Pyannote loading failed: {e}")
|
| 450 |
return None
|
| 451 |
|
|
|
|
| 397 |
except Exception as e:
|
| 398 |
logger.warning(f"Error saving cache for {model_key}: {e}")
|
| 399 |
|
| 400 |
+
def load_pyannote_pipeline(self) -> Optional[Pipeline]:
|
| 401 |
"""Load pyannote speaker diarization pipeline with container-safe settings."""
|
| 402 |
try:
|
| 403 |
console.print(f"[yellow]Loading pyannote.audio pipeline...[/yellow]")
|
| 404 |
|
| 405 |
+
# Fix ONNX Runtime libraries first
|
| 406 |
+
try:
|
| 407 |
+
import subprocess
|
| 408 |
+
subprocess.run([
|
| 409 |
+
'find', '/usr/local/lib/python*/site-packages/onnxruntime',
|
| 410 |
+
'-name', '*.so', '-exec', 'execstack', '-c', '{}', ';'
|
| 411 |
+
], capture_output=True, timeout=10, stderr=subprocess.DEVNULL)
|
| 412 |
+
except:
|
| 413 |
+
pass
|
| 414 |
+
|
| 415 |
# Check for HuggingFace token
|
| 416 |
+
hf_token = os.getenv('HUGGINGFACE_TOKEN') or os.getenv('HF_TOKEN')
|
| 417 |
if not hf_token:
|
| 418 |
console.print("[red]Warning: HUGGINGFACE_TOKEN not found. Some models may not be accessible.[/red]")
|
| 419 |
|
| 420 |
+
# Suppress all warnings during pipeline loading
|
| 421 |
+
import warnings
|
| 422 |
+
import logging
|
| 423 |
+
|
| 424 |
+
# Temporarily disable all warnings and logging
|
| 425 |
+
old_warning_filters = warnings.filters[:]
|
| 426 |
+
warnings.filterwarnings("ignore")
|
| 427 |
+
|
| 428 |
+
# Disable ONNX Runtime logging
|
| 429 |
+
os.environ['ORT_LOGGING_LEVEL'] = '3' # ERROR only
|
| 430 |
+
|
| 431 |
+
# Disable other verbose logging
|
| 432 |
+
logging.getLogger('onnxruntime').setLevel(logging.ERROR)
|
| 433 |
+
logging.getLogger('transformers').setLevel(logging.ERROR)
|
| 434 |
+
|
| 435 |
+
try:
|
| 436 |
pipeline = Pipeline.from_pretrained(
|
| 437 |
"pyannote/speaker-diarization-3.1",
|
| 438 |
use_auth_token=hf_token,
|
| 439 |
cache_dir=str(self.cache_dir / "pyannote")
|
| 440 |
)
|
| 441 |
|
| 442 |
+
# Force CPU execution for all models in pipeline
|
| 443 |
if hasattr(pipeline, '_models'):
|
| 444 |
for model_name, model in pipeline._models.items():
|
| 445 |
if hasattr(model, 'to'):
|
| 446 |
model.to('cpu')
|
| 447 |
+
|
| 448 |
+
console.print(f"[green]SUCCESS: pyannote.audio pipeline loaded successfully on CPU[/green]")
|
| 449 |
+
return pipeline
|
| 450 |
+
|
| 451 |
+
finally:
|
| 452 |
+
# Restore warning filters
|
| 453 |
+
warnings.filters[:] = old_warning_filters
|
| 454 |
|
| 455 |
except Exception as e:
|
| 456 |
+
error_msg = str(e).lower()
|
| 457 |
+
if "executable stack" in error_msg or "onnxruntime" in error_msg:
|
| 458 |
+
console.print("[yellow]ONNX Runtime container warning (attempting workaround)...[/yellow]")
|
| 459 |
+
|
| 460 |
+
# Try alternative approach - load without ONNX-dependent components
|
| 461 |
try:
|
| 462 |
+
# Try loading with CPU-only execution providers
|
| 463 |
+
import onnxruntime as ort
|
| 464 |
+
ort.set_default_logger_severity(4) # FATAL only
|
| 465 |
+
|
| 466 |
+
pipeline = Pipeline.from_pretrained(
|
| 467 |
+
"pyannote/speaker-diarization-3.1",
|
| 468 |
+
use_auth_token=hf_token,
|
| 469 |
+
cache_dir=str(self.cache_dir / "pyannote")
|
| 470 |
+
)
|
| 471 |
+
console.print(f"[green]SUCCESS: pyannote.audio loaded with workaround[/green]")
|
| 472 |
+
return pipeline
|
| 473 |
+
|
| 474 |
+
except Exception as e2:
|
| 475 |
+
console.print(f"[red]ERROR: All pyannote loading methods failed: {e2}[/red]")
|
| 476 |
+
else:
|
| 477 |
+
console.print(f"[red]ERROR: Failed to load pyannote.audio pipeline: {e}[/red]")
|
| 478 |
|
|
|
|
| 479 |
logger.error(f"Pyannote loading failed: {e}")
|
| 480 |
return None
|
| 481 |
|
startup.py
CHANGED
|
@@ -70,15 +70,33 @@ def preload_models():
|
|
| 70 |
import model_preloader
|
| 71 |
logger.info('✅ Model preloader module found')
|
| 72 |
|
| 73 |
-
# Set environment variables
|
| 74 |
env = os.environ.copy()
|
| 75 |
env.update({
|
| 76 |
'ORT_DYLIB_DEFAULT_OPTIONS': 'DisableExecutablePageAllocator=1',
|
| 77 |
'ONNXRUNTIME_EXECUTION_PROVIDERS': 'CPUExecutionProvider',
|
|
|
|
| 78 |
'TF_ENABLE_ONEDNN_OPTS': '0',
|
| 79 |
-
'OMP_NUM_THREADS': '1'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 80 |
})
|
| 81 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 82 |
# Try to run the preloader
|
| 83 |
result = subprocess.run(
|
| 84 |
['python', 'model_preloader.py'],
|
|
@@ -96,18 +114,22 @@ def preload_models():
|
|
| 96 |
else:
|
| 97 |
logger.warning(f'⚠️ Model preloading failed with return code {result.returncode}')
|
| 98 |
if result.stderr:
|
| 99 |
-
#
|
| 100 |
-
|
| 101 |
-
|
| 102 |
-
|
| 103 |
-
|
|
|
|
|
|
|
|
|
|
| 104 |
return False
|
| 105 |
|
| 106 |
except subprocess.TimeoutExpired:
|
| 107 |
logger.warning('⚠️ Model preloading timed out, continuing...')
|
| 108 |
return False
|
| 109 |
except Exception as e:
|
| 110 |
-
|
|
|
|
| 111 |
return False
|
| 112 |
|
| 113 |
def start_web_app():
|
|
|
|
| 70 |
import model_preloader
|
| 71 |
logger.info('✅ Model preloader module found')
|
| 72 |
|
| 73 |
+
# Set comprehensive environment variables for ONNX Runtime
|
| 74 |
env = os.environ.copy()
|
| 75 |
env.update({
|
| 76 |
'ORT_DYLIB_DEFAULT_OPTIONS': 'DisableExecutablePageAllocator=1',
|
| 77 |
'ONNXRUNTIME_EXECUTION_PROVIDERS': 'CPUExecutionProvider',
|
| 78 |
+
'ORT_DISABLE_TLS_ARENA': '1',
|
| 79 |
'TF_ENABLE_ONEDNN_OPTS': '0',
|
| 80 |
+
'OMP_NUM_THREADS': '1',
|
| 81 |
+
'MKL_NUM_THREADS': '1',
|
| 82 |
+
'NUMBA_NUM_THREADS': '1',
|
| 83 |
+
'TOKENIZERS_PARALLELISM': 'false',
|
| 84 |
+
'MALLOC_ARENA_MAX': '2',
|
| 85 |
+
# Additional ONNX Runtime fixes
|
| 86 |
+
'ONNXRUNTIME_LOG_SEVERITY_LEVEL': '3',
|
| 87 |
+
'ORT_LOGGING_LEVEL': 'WARNING'
|
| 88 |
})
|
| 89 |
|
| 90 |
+
# Try to fix ONNX Runtime libraries before running preloader
|
| 91 |
+
try:
|
| 92 |
+
import subprocess
|
| 93 |
+
subprocess.run([
|
| 94 |
+
'find', '/usr/local/lib/python*/site-packages/onnxruntime',
|
| 95 |
+
'-name', '*.so', '-exec', 'execstack', '-c', '{}', ';'
|
| 96 |
+
], capture_output=True, timeout=30)
|
| 97 |
+
except:
|
| 98 |
+
pass # Continue if execstack fix fails
|
| 99 |
+
|
| 100 |
# Try to run the preloader
|
| 101 |
result = subprocess.run(
|
| 102 |
['python', 'model_preloader.py'],
|
|
|
|
| 114 |
else:
|
| 115 |
logger.warning(f'⚠️ Model preloading failed with return code {result.returncode}')
|
| 116 |
if result.stderr:
|
| 117 |
+
# Filter out expected ONNX warnings
|
| 118 |
+
stderr_lines = result.stderr.split('\n')
|
| 119 |
+
important_errors = [line for line in stderr_lines
|
| 120 |
+
if 'executable stack' not in line.lower()
|
| 121 |
+
and 'onnxruntime' not in line.lower()
|
| 122 |
+
and line.strip()]
|
| 123 |
+
if important_errors:
|
| 124 |
+
logger.warning(f'Important errors: {important_errors[:3]}')
|
| 125 |
return False
|
| 126 |
|
| 127 |
except subprocess.TimeoutExpired:
|
| 128 |
logger.warning('⚠️ Model preloading timed out, continuing...')
|
| 129 |
return False
|
| 130 |
except Exception as e:
|
| 131 |
+
if 'executable stack' not in str(e).lower():
|
| 132 |
+
logger.warning(f'⚠️ Model preloading failed: {e}')
|
| 133 |
return False
|
| 134 |
|
| 135 |
def start_web_app():
|