| <script lang="ts"> |
| import { onMount } from "svelte"; |
| import type { I18nFormatter } from "@gradio/utils"; |
| import { createEventDispatcher } from "svelte"; |
| import WaveSurfer from "wavesurfer.js"; |
| import { skip_audio, process_audio } from "../shared/utils"; |
| import WSRecord from "wavesurfer.js/dist/plugins/record.js"; |
| import WaveformControls from "../shared/WaveformControls.svelte"; |
| import WaveformRecordControls from "../shared/WaveformRecordControls.svelte"; |
| import RecordPlugin from "wavesurfer.js/dist/plugins/record.js"; |
| import type { WaveformOptions } from "../shared/types"; |
| import { format_time } from "@gradio/utils"; |
| |
| export let mode: string; |
| export let i18n: I18nFormatter; |
| export let dispatch_blob: ( |
| blobs: Uint8Array[] | Blob[], |
| event: "stream" | "change" | "stop_recording" |
| ) => Promise<void> | undefined; |
| export let waveform_settings: Record<string, any>; |
| export let waveform_options: WaveformOptions = { |
| show_recording_waveform: true |
| }; |
| export let handle_reset_value: () => void; |
| export let editable = true; |
| export let recording = false; |
| |
| let micWaveform: WaveSurfer; |
| let recordingWaveform: WaveSurfer; |
| let playing = false; |
| |
| let recordingContainer: HTMLDivElement; |
| let microphoneContainer: HTMLDivElement; |
| |
| let record: WSRecord; |
| let recordedAudio: string | null = null; |
| |
| |
| let timeRef: HTMLTimeElement; |
| let durationRef: HTMLTimeElement; |
| let audio_duration: number; |
| let seconds = 0; |
| let interval: NodeJS.Timeout; |
| let timing = false; |
| |
| let trimDuration = 0; |
| |
| const start_interval = (): void => { |
| clearInterval(interval); |
| interval = setInterval(() => { |
| seconds++; |
| }, 1000); |
| }; |
| |
| const dispatch = createEventDispatcher<{ |
| start_recording: undefined; |
| pause_recording: undefined; |
| stop_recording: undefined; |
| stop: undefined; |
| play: undefined; |
| pause: undefined; |
| end: undefined; |
| edit: undefined; |
| }>(); |
| |
| function record_start_callback(): void { |
| start_interval(); |
| timing = true; |
| dispatch("start_recording"); |
| if (waveform_options.show_recording_waveform) { |
| let waveformCanvas = microphoneContainer; |
| if (waveformCanvas) waveformCanvas.style.display = "block"; |
| } |
| } |
| |
| async function record_end_callback(blob: Blob): Promise<void> { |
| seconds = 0; |
| timing = false; |
| clearInterval(interval); |
| try { |
| const array_buffer = await blob.arrayBuffer(); |
| const context = new AudioContext({ |
| sampleRate: waveform_settings.sampleRate |
| }); |
| const audio_buffer = await context.decodeAudioData(array_buffer); |
| |
| if (audio_buffer) |
| await process_audio(audio_buffer).then(async (audio: Uint8Array) => { |
| await dispatch_blob([audio], "change"); |
| await dispatch_blob([audio], "stop_recording"); |
| }); |
| } catch (e) { |
| console.error(e); |
| } |
| } |
| |
| $: record?.on("record-resume", () => { |
| start_interval(); |
| }); |
| |
| $: recordingWaveform?.on("decode", (duration: any) => { |
| audio_duration = duration; |
| durationRef && (durationRef.textContent = format_time(duration)); |
| }); |
| |
| $: recordingWaveform?.on( |
| "timeupdate", |
| (currentTime: any) => |
| timeRef && (timeRef.textContent = format_time(currentTime)) |
| ); |
| |
| $: recordingWaveform?.on("pause", () => { |
| dispatch("pause"); |
| playing = false; |
| }); |
| |
| $: recordingWaveform?.on("play", () => { |
| dispatch("play"); |
| playing = true; |
| }); |
| |
| $: recordingWaveform?.on("finish", () => { |
| dispatch("stop"); |
| playing = false; |
| }); |
| |
| let record_mounted = false; |
| |
| const create_mic_waveform = (): void => { |
| if (microphoneContainer) microphoneContainer.innerHTML = ""; |
| if (micWaveform !== undefined) micWaveform.destroy(); |
| if (!microphoneContainer) return; |
| micWaveform = WaveSurfer.create({ |
| ...waveform_settings, |
| normalize: false, |
| container: microphoneContainer |
| }); |
| |
| record = micWaveform.registerPlugin(RecordPlugin.create()); |
| record?.on("record-end", record_end_callback); |
| record?.on("record-start", record_start_callback); |
| record?.on("record-pause", () => { |
| dispatch("pause_recording"); |
| clearInterval(interval); |
| }); |
| |
| record?.on("record-end", (blob) => { |
| recordedAudio = URL.createObjectURL(blob); |
| |
| const microphone = microphoneContainer; |
| const recording = recordingContainer; |
| |
| if (microphone) microphone.style.display = "none"; |
| if (recording && recordedAudio) { |
| recording.innerHTML = ""; |
| create_recording_waveform(); |
| } |
| }); |
| record_mounted = true; |
| }; |
| |
| const create_recording_waveform = (): void => { |
| let recording = recordingContainer; |
| if (!recordedAudio || !recording) return; |
| recordingWaveform = WaveSurfer.create({ |
| container: recording, |
| url: recordedAudio, |
| ...waveform_settings |
| }); |
| }; |
| |
| const handle_trim_audio = async ( |
| start: number, |
| end: number |
| ): Promise<void> => { |
| mode = "edit"; |
| const decodedData = recordingWaveform.getDecodedData(); |
| if (decodedData) |
| await process_audio(decodedData, start, end).then( |
| async (trimmedAudio: Uint8Array) => { |
| await dispatch_blob([trimmedAudio], "change"); |
| await dispatch_blob([trimmedAudio], "stop_recording"); |
| recordingWaveform.destroy(); |
| create_recording_waveform(); |
| } |
| ); |
| dispatch("edit"); |
| }; |
| |
| onMount(() => { |
| create_mic_waveform(); |
| |
| window.addEventListener("keydown", (e) => { |
| const is_focused_in_waveform = |
| recordingContainer && |
| recordingContainer.contains(document.activeElement); |
| if (!is_focused_in_waveform) return; |
| if (e.key === "ArrowRight") { |
| skip_audio(recordingWaveform, 0.1); |
| } else if (e.key === "ArrowLeft") { |
| skip_audio(recordingWaveform, -0.1); |
| } |
| }); |
| }); |
| </script> |
| |
| <div class="component-wrapper"> |
| <div |
| class="microphone" |
| bind:this={microphoneContainer} |
| data-testid="microphone-waveform" |
| /> |
| <div bind:this={recordingContainer} data-testid="recording-waveform" /> |
| |
| {#if (timing || recordedAudio) && waveform_options.show_recording_waveform} |
| <div class="timestamps"> |
| <time bind:this={timeRef} class="time">0:00</time> |
| <div> |
| {#if mode === "edit" && trimDuration > 0} |
| <time class="trim-duration">{format_time(trimDuration)}</time> |
| {/if} |
| {#if timing} |
| <time class="duration">{format_time(seconds)}</time> |
| {:else} |
| <time bind:this={durationRef} class="duration">0:00</time> |
| {/if} |
| </div> |
| </div> |
| {/if} |
| |
| {#if microphoneContainer && !recordedAudio && record_mounted} |
| <WaveformRecordControls |
| bind:record |
| {i18n} |
| {timing} |
| {recording} |
| show_recording_waveform={waveform_options.show_recording_waveform} |
| record_time={format_time(seconds)} |
| /> |
| {/if} |
| |
| {#if recordingWaveform && recordedAudio} |
| <WaveformControls |
| bind:waveform={recordingWaveform} |
| container={recordingContainer} |
| {playing} |
| {audio_duration} |
| {i18n} |
| {editable} |
| interactive={true} |
| {handle_trim_audio} |
| bind:trimDuration |
| bind:mode |
| show_redo |
| {handle_reset_value} |
| {waveform_options} |
| /> |
| {/if} |
| </div> |
| |
| <style> |
| .microphone { |
| width: 100%; |
| display: none; |
| } |
| |
| .component-wrapper { |
| padding: var(--size-3); |
| width: 100%; |
| } |
| |
| .timestamps { |
| display: flex; |
| justify-content: space-between; |
| align-items: center; |
| width: 100%; |
| padding: var(--size-1) 0; |
| margin: var(--spacing-md) 0; |
| } |
| |
| .time { |
| color: var(--neutral-400); |
| } |
| |
| .duration { |
| color: var(--neutral-400); |
| } |
| |
| .trim-duration { |
| color: var(--color-accent); |
| margin-right: var(--spacing-sm); |
| } |
| </style> |
| |