Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -12,7 +12,7 @@ if not IS_DUPLICATE:
|
|
| 12 |
import misaki
|
| 13 |
print('DEBUG', kokoro.__version__, CUDA_AVAILABLE, misaki.__version__)
|
| 14 |
|
| 15 |
-
CHAR_LIMIT = None if IS_DUPLICATE else
|
| 16 |
models = {gpu: KModel().to('cuda' if gpu else 'cpu').eval() for gpu in [False] + ([True] if CUDA_AVAILABLE else [])}
|
| 17 |
pipelines = {lang_code: KPipeline(lang_code=lang_code, model=False) for lang_code in 'ab'}
|
| 18 |
pipelines['a'].g2p.lexicon.golds['qhash'] = 'kˈOkəɹO'
|
|
@@ -140,7 +140,7 @@ with gr.Blocks() as generate_tab:
|
|
| 140 |
out_audio = gr.Audio(label='Output Audio', interactive=False, streaming=False, autoplay=True)
|
| 141 |
generate_btn = gr.Button('Generate', variant='primary')
|
| 142 |
with gr.Accordion('Output Tokens', open=True):
|
| 143 |
-
out_ps = gr.Textbox(interactive=False, show_label=False, info='Tokens used to generate the audio, up to
|
| 144 |
tokenize_btn = gr.Button('Tokenize', variant='secondary')
|
| 145 |
gr.Markdown(TOKEN_NOTE)
|
| 146 |
predict_btn = gr.Button('Predict', variant='secondary', visible=False)
|
|
@@ -172,7 +172,7 @@ with gr.Blocks() as app:
|
|
| 172 |
gr.Markdown(BANNER_TEXT, container=True)
|
| 173 |
with gr.Row():
|
| 174 |
with gr.Column():
|
| 175 |
-
text = gr.Textbox(label='Input Text', info=f"Up to ~
|
| 176 |
with gr.Row():
|
| 177 |
voice = gr.Dropdown(list(CHOICES.items()), value='af_heart', label='Voice', info='Quality and availability vary by language')
|
| 178 |
use_gpu = gr.Dropdown(
|
|
|
|
| 12 |
import misaki
|
| 13 |
print('DEBUG', kokoro.__version__, CUDA_AVAILABLE, misaki.__version__)
|
| 14 |
|
| 15 |
+
CHAR_LIMIT = None if IS_DUPLICATE else 50000
|
| 16 |
models = {gpu: KModel().to('cuda' if gpu else 'cpu').eval() for gpu in [False] + ([True] if CUDA_AVAILABLE else [])}
|
| 17 |
pipelines = {lang_code: KPipeline(lang_code=lang_code, model=False) for lang_code in 'ab'}
|
| 18 |
pipelines['a'].g2p.lexicon.golds['qhash'] = 'kˈOkəɹO'
|
|
|
|
| 140 |
out_audio = gr.Audio(label='Output Audio', interactive=False, streaming=False, autoplay=True)
|
| 141 |
generate_btn = gr.Button('Generate', variant='primary')
|
| 142 |
with gr.Accordion('Output Tokens', open=True):
|
| 143 |
+
out_ps = gr.Textbox(interactive=False, show_label=False, info='Tokens used to generate the audio, up to 5010 context length.')
|
| 144 |
tokenize_btn = gr.Button('Tokenize', variant='secondary')
|
| 145 |
gr.Markdown(TOKEN_NOTE)
|
| 146 |
predict_btn = gr.Button('Predict', variant='secondary', visible=False)
|
|
|
|
| 172 |
gr.Markdown(BANNER_TEXT, container=True)
|
| 173 |
with gr.Row():
|
| 174 |
with gr.Column():
|
| 175 |
+
text = gr.Textbox(label='Input Text', info=f"Up to ~5000 characters per Generate, or {'∞' if CHAR_LIMIT is None else CHAR_LIMIT} characters per Stream")
|
| 176 |
with gr.Row():
|
| 177 |
voice = gr.Dropdown(list(CHOICES.items()), value='af_heart', label='Voice', info='Quality and availability vary by language')
|
| 178 |
use_gpu = gr.Dropdown(
|