--- license: apache-2.0 task_categories: - visual-question-answering tags: - mobile-ui - gui-automation - benchmark - QA testing - tap-accuracy --- # UI-TapBench ## 📝 Summary & Intention UI-TapBench is an open-source benchmark created to evaluate the spatial precision of Large Multimodal Models (LMMs) in mobile environments. As AI agents move toward _"Actionable AI,"_ the ability to translate a natural language instruction into exact screen coordinates is the most common point of failure. This dataset provides a standardized way to measure and improve how models handle dense UI layouts and list-based navigation, ensuring tap reliability in autonomous agents. --- ## 🚀 About Drizz > Reimagining Mobile App Testing with Vision AI. At [Drizz](https://drizz.dev), we're building the world's fastest AI-powered testing agent for mobile apps — no locators, no scripting, just plain English. Mobile teams today move fast, but testing tools haven't kept up. Drizz replaces brittle, locator-based frameworks with a vision-based AI engine that understands your app like a human. With Drizz, teams achieve: - ⚡ 10x Faster Test Cycles - 🎯 97%+ Test Accuracy - 🛡️ Zero Flaky Tests via our vision-based engine We are releasing UI-TapBench to help the community move toward a world where UI automation is as simple, reliable, and _"human-like"_ as possible. --- ## 📊 Dataset Structure Each entry in `metadata.jsonl` follows this schema: | Key | Description | | ------------ | -------------------------------------------------------------- | | `id` | Unique identifier for the sample. | | `image` | Relative path to the screenshot (e.g., `images/841.png`). | | `task` | The natural language command (e.g., _"Tap on second option"_). | | `bbox` | Ground truth coordinates: `[xmin, ymin, xmax, ymax]`. | | `app_name` | The package name of the app being tested. | | `function` | The targeted action type (default: `tap_call_llm`). | ### Example Entry ```json { "id": 841, "image": "images/841.png", "task": "Tap on second option in the list.", "bbox": [42, 733, 1038, 901], "app_name": "com.duolingo", "function": "tap_call_llm" } ``` --- ## 📈 Benchmark Results We evaluated UI-TapBench across leading Large Multimodal Models (LMMs) to measure tap accuracy, spatial precision, and reliability for mobile UI interactions. ### 🔍 Competitor Comparison | Model | Accuracy | Precision | Recall | F1 Score | | ----------------------------------------------------------- | ------------------------------------------------: | ------------------------------------------------: | ------------------------------------------------: | ------------------------------------------------: | | 🏆 Drizz (ours) | 94.51 | 96.22 | 98.16 | 97.18 | | gpt-5.1 | 21.72 | 23.35 | 75.61 | 35.68 | | gpt-5.2 | 44.83 | 45.71 | 95.88 | 61.91 | | gemini-pro | 89.84 | 91.28 | 98.28 | 94.65 | | gemini-flash | 81.44 | 83.78 | 96.67 | 89.77 | | qwen3.5-27b | 92.98 | 94.98 | 97.61 | 96.28 | ### 💡 Key Takeaway The results show that while several models perform well on general UI grounding tasks, Drizz demonstrates the highest benchmark performance on UI-TapBench, achieving strong spatial precision and reliable tap execution even in dense mobile UI layouts. --- ## 📜 License Released under the Apache 2.0 License.