Keywords

1 Introduction

As of 2013, 43 % of adult Americans over age 16 own a tablet or e-reader and 56 % of adult Americans own a smartphone [7, 8], and touch-screen input has rapidly become the primary way many users interact with these mobile computing devices. This trend is especially true for children of pre-school and elementary school age (ages 3 to 10) [5]. Previous work has determined that touch and stroke gesture interaction behaviors differ between adults and children in ways that significantly impact gesture recognition and processing of users’ touch input [2, 3, 6], e.g., children’s touches are more likely to be just outside the boundaries of their intended target [3] and their gestures are less likely to be recognized correctly than adults’ gestures [2, 3].

If touch-screen interfaces were able to determine on the fly whether the user is a child or an adult, applications could switch ad hoc to processing or recognition algorithms tailored for one age group or the other. Such adaptive touch interaction mechanisms would assist users in their touch input, making them more accurate and more successful during touch-screen interaction. Unfortunately, research on age-detection classification is sparse, and what work exists so far focuses on gesture input [6], which provides a far richer source of classification features than simple taps; see for instance Blagojevic et al. [4] for an examination of 114 such features. On the other hand, touch input on its own is more challenging to interpret, because of the rudimentary features provided by current touch-screen hardware, typically limited to touch coordinates and timestamps and, in the best-case scenario, to rough approximations of touch pressure and area size.

We present in this work a technique for classifying users’ age group, i.e., child or adult, from touch input alone. In this work, a child is a person having 6 years at most; at this age, children’s touch input behaviors are the most different from those of adults knowing that children’s input performance and touch accuracy improve with age [3]. Our technique, using a Bayes’ rule classifier, delivered 86.5 % user-independent accuracy on a dataset of 119 participants (89 children ages 3 to 6) when classifying each touch event one at a time and 99 % accuracy with a window of 7+ consecutive touch events. These results establish that it is possible to classify a user as a child or adult with high accuracy using only minimal data provided by all existing touch-screen hardware. Our contributions will inform the next generation of adaptive user interfaces for touch-screen devices that will be able to respond appropriately and on the fly to their users, without pre-configuration or the need of specialized sensors.

2 Touch Input Analysis to Inform Classifier Design

We analyze in this section adults’ and small children’s touch input patterns using two measures that can be readily computed for any touch-screen device, i.e., (1) the time between the moment when the user’s finger touches the screen and the moment when the finger lifts off (Tap-Time) and (2) the distance between the actual touch point where the user’s finger was placed and the center of the touch target (Offset-Distance); see Fig. 1 for an illustration of these measures. We generate frequency distributions of adults’ and children’s tap times and offset distances (see Fig. 2, next page) using the touch dataset of Vatavu et al. [9]. This dataset contains 587 samples from 119 participants (89 children with ages between 3 and 6 years old and 30 young adults).Footnote 1

Fig. 1.
figure 1

Tap-Time (\( \tau \)) and Offset-Distance (\( d \)) touch measures.

Fig. 2.
figure 2

Distribution of the Child and Adult classes in the Tap-Time × Offset-Distance space (left) and frequency histograms of each feature (right) with normal distributions superimposed. Note: we use log e \( - \) transformed Tap-Time and Offset-Distance; a star symbol () indicates that the Kolmogorov-Smirnov test found no significant difference (\( p > .01 \)) between our log e -transformed data and a normal distribution \( N(\mu ,\sigma^{2} ) \) with same mean and standard deviation.

Figure 2 plots the Child and Adult classes in the Tap-Time × Offset-Distance space and the frequency distributions of the two measures. We found that all 4 distributions deviated significantly from normality, so we applied a log e transform on our data. Follow-up Kolmogorov-Smirnov tests confirmed the normality of 3 out of the 4 log e transformed distributions (i.e., Tap-Time for Child, Offset-Distance for Adult, and Offset-Distance for Child, which are marked with a star symbol in Fig. 2). The only distribution that still deviated significantly from normality was Tap-Time for the Adult class (\( D_{(149)} = 0.171, p < .01 \)). However, the effect size of the Kolmogorov-Smirnov test was considerably smaller than Cohen’s suggested limit for small effect sizes (\( r = .014 < .100 \)), while both skewness and kurtosis were close to zero \( (s = - 0.62 \), CI95% \( = [ - 0.81, - 0.42] \) and \( k = - 0.76 \), CI95% \( = \left[ { - 1.16, - 0.36} \right] \)). These results support modeling log e (Tap-Time) and log e (Offset-Distance) data with normal distributions, which in turn leverage the option of designing a Bayes’ rule classifier to discriminate between Child and Adult classes.

Bayes’ rule assigns a candidate measurement \( x \) to class \( j \) for which the a posteriori probability \( p\left( {class_{j} |x} \right) \) is maximized [11] (p. 7):

$$ p\left( {class_{j} |x} \right) = \frac{{p(x|class_{j} ) \cdot p(class_{j} )}}{p(x)} $$
(1)

knowing the class-conditional density functions \( p(x|class_{j} ) \) (i.e., the probability of a randomly chosen pattern \( x \) to lie with class \( j \)) and the a priori probabilities \( p(class_{j} ) \) (i.e., how likely it is to see a pattern of each class); \( p(x) \) is a normalizing factor that can be omitted when comparing a posteriori probabilities directly. For our problem, we have two classes only (Child and Adult), for which the a priori probabilities can be considered equal and thus can be omitted from Eq. 1. Then, Bayes’ rule becomes:

$$ {\text{Assign measurement}}\,x\,{\text{to class}}\,\left\{ {\begin{array}{ll} {Child} & {{\text{if}}~p\left( {x |Child} \right) > p(x|Adult)} \\ {Adult} & {\text{otherwise}} \\ \end{array} } \right. $$
(2)

With each measure and their combination, there are 3 possible classifier designs:

  1. 1.

    Bayes’ rule for T ap -T ime measurements (i.e., the Tap-Time classifier). In this case, x is the log e \( - \) transformed value of a Tap-Time measurement collected from the user, \( p\left( {x |Child} \right) = \frac{1}{{\sqrt {2\pi } \sigma_{c} }}exp\left\{ { - \frac{{\left( {x - \mu_{c} } \right)^{2} }}{{2\sigma_{c}^{2} }}} \right\} \) and \( p\left( {x |Adult} \right) = \frac{1}{{\sqrt {2\pi } \sigma_{a} }}exp\left\{ { - \frac{{\left( {x - \mu_{a} } \right)^{2} }}{{2\sigma_{a}^{2} }}} \right\} \); \( \mu_{c} \) and \( \sigma_{c} \) are the mean and standard deviation of Tap-Time for the Child class, and \( \mu_{a} \) and \( \sigma_{a} \) for Adults. For instance, \( \mu_{c} = 4.99 ms \) and \( \sigma_{c} = 1.04 ms \); see Fig. 2, top-right (remember that these are log e \( - \) transformed values).

  2. 2.

    Bayes’ rule for O ffset -D istance measurements (Offset-Distance classifier). In this case, x is the log e \( - \) transformed value of an Offset-Distance measurement, \( p(x|Child)\, = \,\frac{1}{{\sqrt {2\pi } \sigma_{c} }}\exp \left\{ { - \frac{{(x - \mu_{c} )^{2} }}{{2\sigma_{c}^{2} }}} \right\} \) and \( p\left( {x |Adult} \right) = \frac{1}{{\sqrt {2\pi } \sigma_{a} }}exp\left\{ { - \frac{{\left( {x - \mu_{a} } \right)^{2} }}{{2\sigma_{a}^{2} }}} \right\} \); \( \mu_{c} \) and \( \sigma_{c} \) are the mean and standard deviation of Offset-Distance for the Child class, and \( \mu_{a} \) and \( \sigma_{a} \) for Adults. For instance, \( \mu_{a} = 0.60 ms \) and \( \sigma_{a} = 0.57 ms \); see Fig. 2, bottom-right (remember that these are log e \( - \) transformed values).

  3. 3.

    Combined T ime & O ffset -D istance rule (i.e., the Time-Offset classifier). In this case, \( x \) is a vector consisting of one Tap-Time and one Offset-Distance measurement, \( p\left( {x |Child} \right) = \frac{1}{{2\pi \sqrt {\left| {\sum_{c} } \right|} }}exp\left\{ { - \frac{1}{2} \left( {x - \mu_{c} } \right)^{T} \sum_{c}^{ - 1} \left( {x - \mu_{c} } \right)} \right\} \), \( p\left( {x |Adult} \right) = \frac{1}{{2\pi \sqrt {\left| {\sum_{a} } \right|} }}exp\left\{ { - \frac{1}{2} \left( {x - \mu_{a} } \right)^{T} \sum_{a}^{ - 1} \left( {x - \mu_{a} } \right)} \right\} \); \( \mu_{c} \) is a vector containing the mean values of Tap-Time and Offset-Distance for the Child class and \( \mu_{a} \) the mean vector for Adults; \( \sum_{c} \) and \( \sum_{a} \) represent the covariance matrices for the two classes. For instance, \( \mu_{c} = \left[ {4.99 , 0.98} \right] \) and \( \sum_{c} = \left[ {\begin{array}{*{20}c} { 1.08} & { - 0.11} \\ { - 0.11} & { 0.50} \\ \end{array} } \right] \), where \( 1.08 \) is the variance (i.e., square standard deviation) of Tap-Time, \( 0.50 \) the variance of Offset-Distance, and \( - 0.11 \) the covariance between Tap-Time and Offset-Distance for the Child class (Fig. 2, right).

3 Experiment #1: Classifying Age Group with One Touch Point Only

We conducted a first experiment to evaluate the accuracy of the Bayes’ rule classifiers to predict age group using measurements from one touch point only. We employed the touch dataset of Vatavu et al. [9], which contains 587 touch samples collected from 119 participants (89 children with ages between 3 and 6 years old and 30 young adults). The experiment was a within-subject design with two independent factors:

  1. 1.

    Classifier, nominal with 3 levels: Tap-Time, Offset-Distance, Time-Offset.

  2. 2.

    The number of training participants P from which we estimated the mean and standard deviation values for Tap-Time and Offset-Distance employed by the class-conditional density functions \( p(x|Child) \) and \( p(x|Adult) \). We varied P between 2 and 30 participants for each class. (P = 10 participants means that data from 5 children and 5 adults was used for training.) For each participant, a maximum of 5 samples are available in the dataset [9], making the size of the training set for estimating \( p(x|Child) \) and \( p\left( {x |Adult} \right) \) between 10 and 150.

We compute one dependent variable, the recognition Accuracy of our classifiers, according to the following user-independent procedure from [10]: (1) for each P, we randomly select P children and P adults for training; (2) we randomly select 1 child and 1 adult for testing from the remaining participants, and we classify one randomly-selected touch point for each; (3) we repeat step 2 for 100 times for each training set, and we repeat step 1 for 100 times (i.e., 100 different training sets). Overall, we report classification results from 15 (number of training participants P) × 100 (repetitions of each P) × 200 (classifications for each training set) = 300,000 classification trials. All tests are user-independent, so different data is used for the training and testing sets.

Friedman’s test showed a significant effect of Classifier on recognition Accuracy (\( \chi_{{\left( {2,N = 1500} \right)}}^{2} = 2304.714, p < .001 \)), and post hoc Wilcoxon signed-rank tests showed significant differences for all classifier pairs (\( p < .001 \)) with medium to large effect sizes (\( r \) between .27 and .61); see Fig. 3. The highest accuracy was delivered by the Time-Offset classifier, which had an average performance of 83.9 % and reached 86.5 % with training data from 15 adults and 15 children. The Tap-Time classifier came second with an average performance of 82.8 % and maximum accuracy of 84.9 %. The Offset-Distance classifier exhibited the lowest performance with only 63.8 % accuracy for the maximum number of P = 30 training participants. Friedman’s tests showed significant effects of P over all classifiers (\( p < .001 \)), with Accuracy increasing considerably for both Time-Offset and Tap-Time from P = 2 to 8 participants (71 % to 83 %), after which it continued to increase slowly up to 86.5 %.

Fig. 3.
figure 3

Recognition accuracy of Bayes’ rule classifiers. Notes: P = 10 participants means that touch data from 5 children and 5 adults was used for training. Five touch measurements were used for each training participant. Error bars show 95 % CIs.

4 Experiment #2: Classifying Age Group with a Touch Window

The results from the previous section show good potential for two Bayes’ rule classifiers to discriminate between Child and Adult classes, with accuracy up to 86.5 %. In this section, we show how this accuracy rate can be much improved by employing a majority vote for which the classification decision is taken after analyzing several touch measurements in a row. A touch window represents a sequence of W consecutive touches. We classify each touch with one of our Bayes’ rules and then count the number of Child and Adult votes. The majority count wins the classification and selecting an odd number for W guarantees no ties for our 2-class problem.

We conducted a second experiment to evaluate the accuracy of the Bayes’ rule classifiers to predict users’ age group using measurements from touch windows. The experiment was a within-subject design with three independent factors:

Classifier, nominal with 3 levels: Tap-Time, Offset-Distance, Time-Offset. The number of training participants \( 2 \le P \le 30 \) from which we estimated the mean and standard deviation values for Tap-Time and Offset-Distance employed by the class-conditional density functions \( p(x|Child) \) and \( p(x|Adult) \). The size of the touch window W with 1, 3, 5, 7, 9, 11, 13, and 15 touches.

We compute one dependent variable, the recognition Accuracy of our classifiers, according to the following user-independent procedure [10]: (1) for each P value, we randomly select P children and P adults for training; (2) for each W, we randomly select W children touch samples and W adult touch samples from the remaining participants, and classify the two W-touch windows as Child or Adult; (3) we repeat step 2 for 100 times for each training set, and we repeat step 1 for 100 times (i.e., 100 different training sets). Overall, we report results from 15 (number of training participants P) × 8 (sizes W of the touch window) × 100 (repetitions of each P) × 200 (classifications for each training set) = 2,400,000 classification trials. Because the dataset has a maximum of 5 touch samples per participant, we simulate larger windows by sampling across participants for this experiment. Since the training set shares no users in common with the testing set (user-independent), this approach is equivalent to off-the-shelf use cases.

Friedman’s test detected a significant effect of Classifier on Accuracy (\( \chi_{{\left( {2,N = 12000} \right)}}^{2} = 18362.182, p < .001 \)) and post hoc Wilcoxon signed-rank tests revealed significant differences between all classifier pairs (at \( p < .001 \)) with medium to large effect sizes (\( r \) between .24 and .60). The Offset-Distance classifier exhibited again the poorest performance (70.9 % average accuracy), much lower than the other two classifiers (93.5 % and 94.2 %, respectively), so we disregard it from subsequent analysis. We found a significant effect of touch window size W on Accuracy (\( \chi_{{\left( {7,N = 1500} \right)}}^{2} = 5975.910, p < .001 \) for the Tap-Time classification rule and \( \chi_{{\left( {7,N = 1500} \right)}}^{2} = 6280.743, p < .001 \) for the Time-Offset rule), with the average Accuracy improving for both Tap-Time and Time-Offset from 83.1 % and 83.9 % for W = 1 up to 97.2 % and 97.5 % for W = 15; see Fig. 4. For each touch window, performance improved with more training participants, e.g., the Time-Offset classifier delivered 97.1 % accuracy with W = 11 and P = 4 participants (2 children + 2 adults), which increased to 99.4 % when training data from P = 30 participants (15 children + 15 adults) was used. Overall, the Time-Offset classification rule delivered significantly better performance than Tap-Time (\( Z = - 38.073, p < .001 \)), with a medium effect size (\( r = .25 \)), while the actual average difference in accuracy was only 1.1 %.

Fig. 4.
figure 4

Recognition accuracy of Bayes’ rule classifiers on touch windows. The effect of number of participants P on accuracy shows an increasing trend overall; we highlight the effect of P for some of the touch windows (W = 7, 11, and 15). For W = 1, see Fig. 3. Error bars show 95 % CIs.

5 Discussion, Conclusions, and Future Work

We showed that distance offset and tap time are enough to classify whether the user is a small child or an adult with reasonable accuracy (86.5 %) using only one touch point and very high accuracy (>99 %) with a window of 7+ touches. Although there are some limitations to our study, as we focused only on children ages 3 to 6 years, accuracy results are promising given the rudimentary data we worked with, i.e., touch coordinates and timestamps alone. These results recommend our technique for discrimination of age groups at much finer granularity. However, we must leave such detailed investigations for future work, as they will most likely need examination of additional features to reach similar levels of accuracy. For example, Fig. 5 shows two such measures, Pressure and Touch-Area for 5 age groups and 185 participants [3, 9]. Because such measurements may not be available for all touch-screen hardware, fine age group discrimination may be restricted to some devices only. While we point the community to these opportunities, the exploration of such tradeoffs is future work.

Fig. 5.
figure 5

Potential touch features for multi-class age group classification that show a descending trend in Offset-Distance and ascending trends in Pressure and Touch-Area. Error bars show ± 1 SD.

Meanwhile, our technique can already be incorporated by practitioners into their designs. By placing our discoveries in the larger context set out by previous work on touch-screen interaction design for children [2, 3, 6, 9], we can already anticipate multiple application opportunities made possible by our classification technique:

  1. (1)

    Adaptive widget layouts. Since we know from the literature that children miss on-screen touch targets more often than adults [3], are less precise [9], and have more difficulty with smaller targets [3], children would benefit from an interface layout with larger widgets and more space in-between. Our technique makes it possible to detect that the user is a child and, therefore, the application can reconfigure the interface layout on the fly.

  2. (2)

    Intelligent widget activation. When changing the widget layout is not desirable, our technique makes it possible to infer the user’s intended target for near-misses by applying a distance offset filter less stringent for children than for adults, e.g., the app may accept farther-away touches and still activate the target if it knows the user is a child. The layout and size of the widgets stay the same, but the application will apply different age-group-dependent processing. This mechanism will provide more flexibility for the application developer by allowing them to use the same interface for multiple target age groups. The increase in touch target acquisition flexibility while maintaining layout appearance is a desirable design option when children and adults use the interface collaboratively.

  3. (3)

    Adaptive activation of gesture processing techniques. Knowing whether the user is a child or an adult can be used to call specialized event handlers in the touch processing chain of the application. For instance, we know that gesture recognizers perform worse on children’s gestures [3], so the application could benefit from the age group knowledge to know when to activate one recognizer or another [4, 10]. Also, we know that visual feedback during gesture input is more important for children than for adults [2], and the app could use the age group data to increase or dim down visual feedback as necessary, which will provide more cues for children and more screen space for adults to display other app widgets and data.

  4. (4)

    Adaptive filtering of content and access to applications. Knowing whether the user is a child or an adult can be used to automatically filter Internet content and to disable access to applications according to parental monitoring settings [1]. However, in contrast to existing approaches, our technique can inform the app right away (from the first touch or just a few touches) that the ownership of the device has changed and can enforce or disable parental settings automatically.

All these application opportunities can be easily incorporated into existing interfaces, increasing adaptiveness to users. We hope that this first examination of age group classification with basic, readily collectable touch data will empower the community with a new tool to make touch interfaces even more adaptable for users, and will inspire researchers to gain richer understanding of touch input for different age groups.