5.9k★by staticai
android-automation – OpenClaw Skill
android-automation is an OpenClaw Skills integration for coding workflows. Control Android devices via ADB with support for UI layout analysis (uiautomator) and visual feedback (screencap). Use when you need to interact with Android apps, perform UI automation, take screenshots, or run complex ADB command sequences.
Skill Snapshot
| name | android-automation |
| description | Control Android devices via ADB with support for UI layout analysis (uiautomator) and visual feedback (screencap). Use when you need to interact with Android apps, perform UI automation, take screenshots, or run complex ADB command sequences. OpenClaw Skills integration. |
| owner | staticai |
| repository | staticai/android-adb |
| language | Markdown |
| license | MIT |
| topics | |
| security | L1 |
| install | openclaw add @staticai/android-adb |
| last updated | Feb 7, 2026 |
Maintainer

name: android-automation description: Control Android devices via ADB with support for UI layout analysis (uiautomator) and visual feedback (screencap). Use when you need to interact with Android apps, perform UI automation, take screenshots, or run complex ADB command sequences.
Android Automation
Control and automate Android devices using ADB, uiautomator, and screencap.
Connecting Devices
USB Connection
- Enable Developer Options and USB Debugging on the device.
- Connect via USB and verify with
adb devices.
Wireless Connection (Android 11+)
- Enable Wireless Debugging in Developer Options.
- Pairing: Find the IP, port, and pairing code in the "Pair device with pairing code" popup.
adb pair <ip>:<pairing_port> <pairing_code> - Connecting: Use the IP and port shown on the main Wireless Debugging screen.
adb connect <ip>:<connection_port> - Verify with
adb devices.
Common Workflows
Launching an App
Use the monkey tool to launch apps by package name:
adb shell monkey -p <package_name> -c android.intent.category.LAUNCHER 1
Analyzing the UI
Dump and pull the UI hierarchy to find coordinates:
adb shell uiautomator dump /sdcard/view.xml && adb pull /sdcard/view.xml ./view.xml
Then grep for text or resource IDs to find bounds="[x1,y1][x2,y2]".
Interacting with Elements
- Tap:
adb shell input tap <x> <y> - Text:
adb shell input text "<text>"(Note: Use%\sfor spaces in some environments or handle quoting carefully) - Keyevent:
adb shell input keyevent <keycode>(Home: 3, Back: 4, Power: 26, Search: 84, Enter: 66) - Swipe:
adb shell input swipe <x1> <y1> <x2> <y2> <duration_ms>
Visual Verification
Take a screenshot to verify the state:
adb shell screencap -p /sdcard/screen.png && adb pull /sdcard/screen.png ./screen.png
Tips
- Search: Use
input keyevent 84to trigger search in many apps. - Wait: Use
sleep <seconds>between commands to allow the UI to update. - Coordinates: Calculate the center of
[x1,y1][x2,y2]for reliable taps.
No README available.
Permissions & Security
Security level L1: Low-risk skills with minimal permissions. Review inputs and outputs before running in production.
Requirements
- OpenClaw CLI installed and configured.
- Language: Markdown
- License: MIT
- Topics:
FAQ
How do I install android-automation?
Run openclaw add @staticai/android-adb in your terminal. This installs android-automation into your OpenClaw Skills catalog.
Does this skill run locally or in the cloud?
OpenClaw Skills execute locally by default. Review the SKILL.md and permissions before running any skill.
Where can I verify the source code?
The source repository is available at https://github.com/openclaw/skills/tree/main/skills/staticai/android-adb. Review commits and README documentation before installing.
