Spanish-F5-API / README.md
eloicito333's picture
add LICENSE file and update README with project details and usage examples
3e56a3f
---
license: mit
---
# Spanish-F5 TTS Inference API (Hugging Face)
This project exposes a Hugging Face Inference Endpoint for Spanish-F5, a Spanish-adapted version of the F5-TTS model. It takes reference audio and a target sentence, and synthesizes speech in the same voice.
> ✨ Live inference is powered by Hugging Face Inference Endpoints.
---
## 🔗 Credit
This project is based on [jpgallegoar/Spanish-F5](https://github.com/jpgallegoar/Spanish-F5).
- Spanish-F5 by [@jpgallegoar](https://github.com/jpgallegoar)
- [Model weights on Hugging Face](https://huggingface.co/jpgallegoar/F5-Spanish/)
Addapted by [@eloicito333](https://github.com/eloicito333).
Licensed under the MIT License.
---
## ⚙️ How It Works
### 🔽 Request Parameters
Send a POST request with a JSON body to the Hugging Face Inference Endpoint:
```json
{
"ref_audio": "<base64-encoded WAV>", // string, required
"ref_text": "Hola, ¿cómo estás?", // string, optional (transcript of ref_audio)
"gen_text": "Estoy muy bien, gracias.", // string, required (text to synthesize)
"remove_silence": true, // boolean, optional (default: true)
"speed": 1.0, // number, optional (default: 1.0)
"cross_fade_duration": 0.15 // number, optional (default: 0.15)
}
```
### 🔼 Response Object
The response will be a JSON object:
```json
{
"success": true, // boolean: true if synthesis succeeded
"audio_base64": "<base64-encoded WAV output>" // string: base64 WAV audio (if success)
}
```
If an error occurs:
```json
{
"success": false,
"error": "TypeError: some descriptive message" // string: error description
}
```
Use the `audio_base64` field to decode and save the resulting audio.
---
## 🤖 Node.js Client Example (Using Fetch)
```js
import "fs"
async function sendAudio() {
const audioBuffer = fs.readFileSync("./example.wav");
const audioBase64 = audioBuffer.toString("base64");
const response = await fetch("https://your-hf-endpoint-url", {
method: "POST",
headers: { "Content-Type": "application/json" },
body: JSON.stringify({
ref_audio: audioBase64,
ref_text: "Hola, ¿cómo estás?",
gen_text: "Estoy muy bien, gracias.",
remove_silence: true,
speed: 1.0,
cross_fade_duration: 0.15,
})
});
const result = await response.json();
if (result.audio_base64) {
fs.writeFileSync("output.wav", Buffer.from(result.audio_base64, "base64"));
console.log("Audio saved to output.wav");
} else {
console.error("Error:", result);
}
}
sendAudio();
```
---
## 🔬 Python Client Example (Optional)
```python
import requests
import base64
with open("ref.wav", "rb") as f:
audio_base64 = base64.b64encode(f.read()).decode("utf-8")
response = requests.post("https://your-hf-endpoint-url", json={
"ref_audio": audio_base64,
"ref_text": "Hola, ¿cómo estás?",
"gen_text": "Estoy muy bien, gracias.",
"remove_silence": True,
"speed": 1.0,
"cross_fade_duration": 0.15
})
if response.ok and response.json().get("audio_base64"):
with open("output.wav", "wb") as out:
out.write(base64.b64decode(response.json()["audio_base64"]))
print("Audio saved to output.wav")
else:
print("Error:", response.json())
```
---
## 🎓 License
MIT License. See [LICENSE](./LICENSE) for more information.
---
## ✏️ Author
Addapted by [@eloicito333](https://github.com/eloicito333).