This page explains how Android processes the various inputs it receives from the keyboard, sensors, and more.
The Android haptics subsystem refers to hardware and software features that contribute to the creation of stimuli through the sense of touch. This section provides guidance and compliance instructions on the best use of Android haptics APIs.
The Android input subsystem nominally consists of an event pipeline that traverses multiple layers of the system. At the lowest layer, the physical input device produces signals that describe state changes such as key presses and touch contact points.
The Android Neural Networks API (NNAPI) runs computationally intensive operations for machine learning. This document provides an overview on how to implement a Neural Networks API driver for Android 9.
Peripherals and accessories
Using a suite of standard protocols, you can implement compelling peripherals and other accessories that extend Android capabilities in a wide range of Android-powered devices.
Android sensors give apps access to a mobile device's underlying
physical sensors. They are data-providing virtual devices defined by
sensors.h, the sensor Hardware Abstraction Layer (HAL).
Context Hub Runtime Environment
Context Hub Runtime Environment (CHRE) provides a common platform for running system-level apps on a low-power processor, with a simple, standardized, embedded-friendly API. CHRE makes it easy for device OEMs to offload processing from the applications processor, to save battery and improve various areas of the user experience, and enable a class of always-on, contextually aware features.