| --- |
| library_name: transformers |
| tags: [] |
| --- |
| # DACTYL Text Detector |
| ## Training Configuration |
| ```json |
| { |
| "training_split": "training", |
| "evaluation_split": "testing", |
| "results_path": "libauc-bert-tiny.csv", |
| "num_epochs": 1, |
| "model_path": "ShantanuT01/dactyl-bert-tiny-pretrained", |
| "tokenizer": "prajjwal1/bert-tiny", |
| "optimizer": "SOTAs", |
| "optimizer_type": "libauc", |
| "optimizer_args": { |
| "lr": 1e-05 |
| }, |
| "loss_fn": "tpAUC_KL_Loss", |
| "reset_classification_head": true, |
| "loss_type": "libauc", |
| "loss_fn_args": { |
| "data_len": 466005 |
| }, |
| "needs_loss_fn_as_parameter": false, |
| "save_path": "ShantanuT01/dactyl-bert-tiny-finetuned", |
| "training_args": { |
| "batch_size": 64, |
| "needs_sampler": true, |
| "needs_index": true, |
| "shuffle": false, |
| "sampling_rate": 0.5, |
| "apply_sigmoid": true |
| }, |
| "best_model_path": "best-tpauc-model-bert-tiny" |
| } |
| ``` |
| ## Results |
| | model | AP Score | AUC Score | OPAUC Score | TPAUC Score | |
| |:--------------------------------------------------------------------------------------|-----------:|------------:|--------------:|--------------:| |
| | DeepSeek-V3 | 0.986177 | 0.997357 | 0.991422 | 0.916469 | |
| | ShantanuT01/fine-tuned-Llama-3.2-1B-Instruct-apollo-mini-RedditWritingPrompts-testing | 0.0242264 | 0.805738 | 0.619715 | 0 | |
| | ShantanuT01/fine-tuned-Llama-3.2-1B-Instruct-apollo-mini-abstracts-testing | 0.473122 | 0.955828 | 0.857774 | 0.0525074 | |
| | ShantanuT01/fine-tuned-Llama-3.2-1B-Instruct-apollo-mini-news-testing | 0.0223995 | 0.776301 | 0.579092 | 0 | |
| | ShantanuT01/fine-tuned-Llama-3.2-1B-Instruct-apollo-mini-reviews-testing | 0.0167409 | 0.861497 | 0.526359 | 0 | |
| | ShantanuT01/fine-tuned-Llama-3.2-1B-Instruct-apollo-mini-student_essays-testing | 0.0102312 | 0.621629 | 0.51884 | 0 | |
| | ShantanuT01/fine-tuned-Llama-3.2-1B-Instruct-apollo-mini-tweets-testing | 0.200169 | 0.949163 | 0.728122 | 0 | |
| | claude-3-5-haiku-20241022 | 0.96485 | 0.993769 | 0.977445 | 0.780686 | |
| | claude-3-5-sonnet-20241022 | 0.981121 | 0.997236 | 0.987999 | 0.883237 | |
| | gemini-1.5-flash | 0.93374 | 0.989888 | 0.955699 | 0.57296 | |
| | gemini-1.5-pro | 0.856575 | 0.96967 | 0.909663 | 0.190672 | |
| | gpt-4o-2024-11-20 | 0.950062 | 0.990051 | 0.968528 | 0.695722 | |
| | gpt-4o-mini | 0.992634 | 0.999149 | 0.995225 | 0.953515 | |
| | llama-3.2-90b | 0.895571 | 0.978201 | 0.933229 | 0.366601 | |
| | llama-3.3-70b | 0.952439 | 0.992252 | 0.968929 | 0.699379 | |
| | mistral-large-latest | 0.979345 | 0.996937 | 0.986758 | 0.871375 | |
| | mistral-small-latest | 0.986836 | 0.997499 | 0.991305 | 0.915284 | |
| | overall | 0.976784 | 0.980759 | 0.947371 | 0.492028 | |
| |