Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -29,3 +29,12 @@ In order to run the code you need to set up the following keys and add them to .
|
|
| 29 |
```bash
|
| 30 |
python run_pipeline.py --file example.txt --model gpt-4-1106-preview
|
| 31 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
```bash
|
| 30 |
python run_pipeline.py --file example.txt --model gpt-4-1106-preview
|
| 31 |
```
|
| 32 |
+
|
| 33 |
+
## Run Gradio app locally
|
| 34 |
+
|
| 35 |
+
```bash
|
| 36 |
+
python app.py
|
| 37 |
+
```
|
| 38 |
+
|
| 39 |
+
## Demo
|
| 40 |
+
Demo available [here](https://huggingface.co/spaces/vladbogo/Filtir)
|
app.py
CHANGED
|
@@ -6,7 +6,7 @@ from run_pipeline import get_fact_checked
|
|
| 6 |
|
| 7 |
def fact_check_function(text, model):
|
| 8 |
# Assume the text is already read from the user input, so we don't need to open a file here
|
| 9 |
-
out = get_fact_checked(text, mode="
|
| 10 |
return out["fact_checked_md"]
|
| 11 |
|
| 12 |
|
|
|
|
| 6 |
|
| 7 |
def fact_check_function(text, model):
|
| 8 |
# Assume the text is already read from the user input, so we don't need to open a file here
|
| 9 |
+
out = get_fact_checked(text, mode="fast", model=model)
|
| 10 |
return out["fact_checked_md"]
|
| 11 |
|
| 12 |
|