File size: 2,731 Bytes
5cd4fb2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9fac9be
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
import altair as alt
import numpy as np
import pandas as pd
import streamlit as st

"""
# Welcome to Streamlit!

Edit `/streamlit_app.py` to customize this app to your heart's desire :heart:.
If you have any questions, checkout our [documentation](https://docs.streamlit.io) and [community
forums](https://discuss.streamlit.io).

In the meantime, below is an example of what you can do with just a few lines of code:
"""

num_points = st.slider("Number of points in spiral", 1, 10000, 1100)
num_turns = st.slider("Number of turns in spiral", 1, 300, 31)

indices = np.linspace(0, 1, num_points)
theta = 2 * np.pi * num_turns * indices
radius = indices

x = radius * np.cos(theta)
y = radius * np.sin(theta)

df = pd.DataFrame({
    "x": x,
    "y": y,
    "idx": indices,
    "rand": np.random.randn(num_points),
})

st.altair_chart(alt.Chart(df, height=700, width=700)
    .mark_point(filled=True)
    .encode(
        x=alt.X("x", axis=None),
        y=alt.Y("y", axis=None),
        color=alt.Color("idx", legend=None, scale=alt.Scale()),
        size=alt.Size("rand", legend=None, scale=alt.Scale(range=[1, 150])),

        
    ))


# import streamlit as st
# import whisper
# import tempfile
# import os
# import torchaudio

# # Title and description
# st.title("🎧 Whisper Audio Transcriber")
# st.markdown("Upload a `.wav` or `.mp3` file to get transcribed text with timestamps using Whisper.")

# # Load Whisper model
# @st.cache_resource
# def load_model():
#     return whisper.load_model("base")

# model = load_model()
# st.success("βœ… Whisper model loaded!")

# # File uploader
# audio_file = st.file_uploader("Upload audio file", type=["wav", "mp3"])

# if audio_file is not None:
#     # Save uploaded file temporarily
#     with tempfile.NamedTemporaryFile(delete=False, suffix=".wav") as tmp_file:
#         tmp_file.write(audio_file.read())
#         temp_path = tmp_file.name

#     # Convert MP3 to WAV if needed
#     if audio_file.name.endswith(".mp3"):
#         waveform, sample_rate = torchaudio.load(temp_path)
#         wav_path = temp_path.replace(".wav", "_converted.wav")
#         torchaudio.save(wav_path, waveform, sample_rate)
#         os.remove(temp_path)
#         temp_path = wav_path

#     # Transcription
#     st.info("πŸ“ Transcribing...")
#     result = model.transcribe(temp_path)

#     # Display segments
#     st.subheader("πŸ•’ Segments with Timestamps")
#     for segment in result["segments"]:
#         st.markdown(f"**[{segment['start']:.2f}s - {segment['end']:.2f}s]**: {segment['text']}")

#     # Full transcription
#     st.subheader("🧾 Full Transcript")
#     st.text_area("Transcribed Text", result["text"], height=250)

#     # Clean up
#     os.remove(temp_path)