Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Machine learning based canine posture estimation using inertial data

Abstract

The aim of this study was to design a new canine posture estimation system specifically for working dogs. The system was composed of Inertial Measurement Units (IMUs) that are commercially available, and a supervised learning algorithm which was developed for different behaviours. Three IMUs, each containing a 3-axis accelerometer, gyroscope, and magnetometer, were attached to the dogs’ chest, back, and neck. To build and test the model, data were collected during a video-recorded behaviour test where the trainee assistance dogs performed static postures (standing, sitting, lying down) and dynamic activities (walking, body shake). Advanced feature extraction techniques were employed for the first time in this field, including statistical, temporal, and spectral methods. The most important features for posture prediction were chosen using Select K Best with ANOVA F-value. The individual contributions of each IMU, sensor, and feature type were analysed using Select K Best scores and Random Forest feature importance. Results showed that the back and chest IMUs were more important than the neck IMU, and the accelerometers were more important than the gyroscopes. The addition of IMUs to the chest and back of dog harnesses is recommended to improve performance. Additionally, statistical and temporal feature domains were more important than spectral feature domains. Three novel cascade arrangements of Random Forest and Isolation Forest were fitted to the dataset. The best classifier achieved an f1-macro of 0.83 and an f1-weighted of 0.90 for the prediction of the five postures, demonstrating a better performance than previous studies. These results were attributed to the data collection methodology (number of subjects and observations, multiple IMUs, use of common working dog breeds) and novel machine learning techniques (advanced feature extraction, feature selection and modelling arrangements) employed. The dataset and code used are publicly available on Mendeley Data and GitHub, respectively.

1 Introduction

Animals express their feelings and emotions through behaviour, therefore behavioural monitoring offers the opportunity to deepen our understanding of animal health and well-being. Ethogram markings have been traditionally used to record and qualify a set of behaviours shown in a particular setting for a determined duration, allowing the discovery of patterns, similarities and differences between subjects [1, 2]. Accordingly, researchers and experts typically use ethograms to characterise, quantify and monitor animal behaviour in order to evaluate well-being and temperament [3].

Ethograms are usually designed for each experimental protocol to include relevant behaviours considering the subjects and research questions under investigation. Their implementation normally involves manual annotation performed by experts during the observation session, or afterwards through video analysis of recorded sessions [1]. This method is time-consuming and may require expert knowledge, therefore ethogram markings are generally employed in short sessions with a selected group of subjects.

In order to overcome the challenges related to obtaining individual-level behavioural data on dogs in the medium and long term, smart canine activity monitors have been developed. The recent evolution of motion sensing, processing power and artificial intelligence technology has enabled the development of automated systems for behavioural estimation and monitoring [411]. They not only eliminate the need for direct observation and assessment by trained professionals but also remove the subjectivity and error associated with human markings. Moreover, they allow continuous recording with an unprecedented level of detail.

Machine learning techniques, such as supervised learning algorithms, have been employed to derive a classification model based on the data provided by inertial measurement units (IMUs) [59, 12]. The units are usually composed of a 3-axis accelerometer, gyroscope, and/or magnetometer, providing a reliable, accurate, cost-effective, and energy-efficient solution for motion analysis. Moreover, these sensors are small, lightweight, and low-cost. These have significant advantages when compared to motion recognition systems based on image analysis from camera systems [1316] which pose constraints in terms of mobility relative to the camera and algorithm complexity.

The aim of this study was to develop the first posture estimation system specifically for working dogs to create automatic ethograms considering five canine behaviours (walking, standing, sitting, lying down, and body shake). The data collection was designed to consider the specific application of this system to working dogs. Accordingly, only two common working dog breeds were included and multiple IMUs were positioned to emulate the harness used by dogs.

The data preprocessing technique utilised advanced feature extraction methods for the first time in this field. Three novel machine learning architectures were implemented and evaluated to improve the rate of correct detection of less frequent behaviour. The analysis of feature importance indicated that the back was the most influential sensor position and accelerometers were the most critical sensor type for posture estimation. The classification results achieved in this study demonstrate the superior performance of this model compared to previous research.

The potential applications of canine activity monitoring in the working dog industry which motivated the development of this work are presented in Section 1, while Section 2 examines the state-of-the-art in this area and identifies the research gaps. Finally, Section 3 elaborates on the novelty and highlights of this study.

1.1 Motivation

The COVID-19 pandemic has profoundly changed people’s behaviour [17], and impacted dog-owner affective experiences and relationships [18], which in turn, correlate with dog’s physical activity [19]. Smart canine activity monitors have been developed to accurately and reliably monitor activity and health parameters to indicate overall well-being [9, 20, 21]. In order to achieve that, they need to first quantify activity levels and identify important behaviours. Subsequently, such data could be fed to specific algorithms to extract meaningful information about the targeted application.

This technological advancement provides an enormous opportunity not only for pet owners but also for working dog organisations including those involved in guide, assistance, police, search and rescue dogs, etc. These devices could be used on guide dogs to provide their organisation with well-being information while they are off-site working with their visually-impaired partners, as a means to address the current challenges in assessing and reporting well-being. They could also enable communication between assistance dogs and their partner’s support systems, such as carers and healthcare professionals, as these dogs are trained to perform specific postures and behaviours to assist their partners. This also applies to search and rescue dogs, who are trained to communicate with their handlers through body postures while working remotely [4, 22, 23], and open-field guard or shepherd dogs [7]. In the case of all types of working dogs, they could also assist in creating automatic scoring for the assessment of behavioural performance [24] and computer-canine training system [10, 25, 26]. Activities estimated by a posture recognition algorithm during a behaviour test could be used to predict ethogram items, rater scoring or ultimately the dog’s probability of success in their training programme [24].

This technology enables a range of applications including assessment of well-being and injury recovery. Canine activity monitors could be used for home activity monitoring of pet dogs with chronic conditions especially those affected by orthopaedic, neuromuscular, or neurological illnesses such as osteoarthritis, muscular dystrophy [27] or heart failure [28]. For example, they could provide an objective assessment method of evaluating dogs during an injury recovery process [9] and suffering from osteoarthritis [29, 30]. They could also be used to monitor skin conditions by identifying pruritic behaviours in allergic dogs and the occurrence of seizures in epileptic dogs [31].

Another possible application of canine posture recognition systems regards the monitoring of well-being as some activity and health parameters have been associated with the occurrence of emotional states and stress in dogs [20, 32]. They could also be utilised for noise cancellation in the interpretation of physiological measurements [4, 33]. Furthermore, they could also be used to measure activity at night, as rest was positively associated with success in apprentice guide dogs [34], and static and inactive behaviours were associated with anxiety and hyper-vigilance [35].

1.2 Related work

The work presented in this manuscript builds upon the existing literature on posture recognition systems for dogs. These systems can be classified according to the type of hardware and software technology deployed. The former relates to the method for gathering data for posture classification including inertial (accelerometer and gyroscope) or image (camera) based systems [36]. The latter is responsible for estimating posture based on data from sensors. It can take place in either real-time or non-real-time and be embedded, computer, or cloud-based.

Several classification algorithms featuring statistical models [10], supervised machine learning [5, 6, 9] or knowledge engineering [11] techniques have been deployed to identify key canine postures and activities. In particular, this work features an inertial sensing, non-real-time, computer-based posture classification system.

The possibility of estimating canine posture using IMUs was first investigated by Ribeiro et al. [22] and Brugarolas et al. [25], and subsequent attempts were introduced later [5, 6, 11]. Since then the design of several posture recognition systems has been reported in the literature using both custom-designed [4, 5, 25] and commercial inertial systems [8, 9, 12, 37, 38]. Some canine activity monitoring devices are currently commercially available [21, 35, 39, 40].

Table 1 summarises, in chronological order, nine published studies whose main goal was to develop or evaluate a canine posture estimation system based on inertial sensors. It includes information on subjects, sensors and placement, data collection methods, data pre-processing techniques, classification algorithms, as well as train-test split and results.

thumbnail
Table 1. Canine posture recognition algorithms using IMUs reported in literature.

https://doi.org/10.1371/journal.pone.0286311.t001

Even though most of the studies present overall accuracy as the main performance measure, comparisons across studies must be carefully made. This metric is highly influenced by class imbalance. In other words, accuracy becomes unrepresentative of the algorithm’s performance as the number of observations in each category becomes dissimilar. Moreover, using data from the same subjects or same breeds in the training and test sets causes the performance metric to be higher compared to using different subjects [5, 37] or breeds [7]. Controlling for these factors considering the final model’s intended application is vital in estimating its real-world performance. Several previously reported studies insufficiently reported the number of observations per class and subject, and the criteria used for splitting the original dataset into training and test sets, even though they significantly affect the model’s accuracy.

Some research reports failed to include information regarding important methodological details such as the device sampling rate [11], window size [6] and overlap [5, 7, 8, 11], data collection methods [12], and feature extraction methods [9]. Insufficient information and incompatible methodological designs pose great challenges when comparing the performance achieved in different studies and limit their contribution to state-of-the-art knowledge. Moreover, Kumpulainen et al. [38] was the only study to publicly share the dataset used for model development.

The number of subjects used for developing canine posture recognition algorithms was rather limited in early research; however, larger sample sizes were reported in more recent studies [7, 8, 37, 38]. Gerencser et al. [7] used two breeds of dogs and attained an inter-subject accuracy of 83%, and Kumpulainen et al. [8] used 20 breeds of dogs and achieved an accuracy of 76%. The lower figure in the latter study can be explained as the use of dogs from different breeds introduces more variability negatively impacting that metric. However, it is unclear whether the same subjects were used for training and testing the algorithm, which would positively affect that metric [7, 37]. To address these issues, this study included 41 dogs belonging to common working dog breeds.

Additionally, two studies were found whose aim was to evaluate commercially available canine activity monitors [35, 39]. These studies insufficiently describe the algorithm built as their primary objective was to assess and validate the classification model used in such commercially available devices, therefore they were not included in Table 1. The first one used PetDialog+ powered by Oggi (Tel Aviv, Israel) and Zoetis (Dublin, Ireland) smart collars on 51 dogs classifying 8 different activities including walking, trotting, canter/galloping, sleeping, static/inactive, eating, drinking, and head shaking. It achieved a mean balanced accuracy of 89% and f1-score of 95% [35].

The second study used Whistle smart collars (MacLean, USA) in a large study involving a canine posture training database containing data from over 2,500 dogs. The system was capable of identifying 14 activities, including nine behaviours: drinking, eating, licking object, licking self, petting, rubbing, scratching, and sniffing; and five postures: lying down, sitting, standing, vigorous, and walking. It achieved a mean balanced accuracy of 85% and macro f1-score of 69% [39] using group cross-fold validation.

Importantly, no study was found to develop a posture recognition system specifically for working dogs. This limitation in the state-of-the-art research was addressed in this research by collecting data on common working dog breeds and attaching multiple sensors to a harness similar to the one used by working dogs. None of the classification algorithms presented in the literature experimented with advanced feature extraction methods or employed anomaly detection classifiers to identify minority classes. Hence, these were explored in the present work which also addressed issues identified in previous research. In particular, it employed more robust cross-validation techniques and reported comprehensive performance metrics for a more complete evaluation of the model.

1.3 Novelty and contribution

The main goal of this work was to build a novel system capable of reliably identifying key canine behaviours in working dogs. In addition to the most common dog postures, namely: walking, sitting, standing, and lying down, it was also of interest to predict body shake as it characterises a coping behaviour shown when dogs experience acute stress [42]. This study comprises four main methodological novelties, advancing the state-of-the-art by:

  1. providing the largest open access dataset for dog activity recognition specific to predominant working dog breeds, namely, Labrador Retrievers, Golden Retrievers, and their crosses [4345];
  2. applying advanced feature extraction methods for the first time in this field;
  3. evaluating three novel architectures of machine learning classifiers to address the natural class imbalance, including anomaly detection;
  4. utilising data from multiple IMUs (i.e., located by the neck, back, and chest) and investigating the effect of IMU placement in a large dataset.

With the advent of FAIR principles [46], a significant improvement has been observed as authors adhere to best practices by sharing raw datasets [37, 38, 47]. Accordingly, this work contributes to the state-of-the-art by providing the annotated dataset on Mendeley Data (link) [48]. The code developed to implement the methodology which yielded the results presented in this study was also made publicly available on GitHub (link). This work proposes a framework for future studies to address the methodological issues discussed in the detailed analysis of past studies. We attempt to establish a model to promote the interoperability of research outputs, considering the particular technicalities in the field of dog activity recognition.

2 Methods

2.1 Subjects

The subjects were 42 healthy apprentice dogs participating in the assistance dog programme at the Irish Guide Dogs for the Blind (IGDB)’s Training Centre in Cork, Ireland. The IGDB is a charitable organisation that provides guide and service dogs to help persons who are vision impaired and families of children with autism. The dogs were Golden Retrievers (N = 1, Mean Age = 14.6 months), Labrador Retrievers (N = 16, Mean Age = 17.3 month) and Crosses (N = 25, Mean Age = 11.6 months).

2.2 Ethical approval

The Health Products Regulatory Authority (HPRA) is the competent authority in Ireland responsible for the implementation of EU legislation (Directive 2010/63/EU) for the protection of animals used for scientific purposes. Practices not likely to cause pain, suffering, distress or lasting harm equivalent to, or higher than, that caused by the introduction of a needle in accordance with good veterinary practice fall outside the scope of HPRA Scientific Animal Protection Legislation. Consequently, no special permission from HPRA was required considering that the nature of the present work was non-invasive.

The Animal Ethics Experimentation Committee (AEEC) and Social Research Ethics Committee (SREC) at University College Cork (UCC) reviewed and approved this study’s data collection procedures as described below under request numbers 2019-007 and 2019-016, respectively. All participants were briefed on the experimental protocol and signed the written informed consent sheet.

2.3 Devices and data

Inertial data were gathered during the data collection session by three IMUs GT9X Links (Actigraph, Pensacola, USA) containing a three-axial accelerometer, gyroscope, and magnetometer. ActiLife software (Actigraph, Pensacola, USA) was used to initialise the devices with a 100Hz sampling rate.

The IMUs were attached to fabric straps on the dogs through hook-and-loop fasteners glued to the devices and on the fabric straps (Xsens, Enschede, the Netherlands). The sensors were placed on the dogs’ neck, back, and chest following the placement shown in Fig 1.

thumbnail
Fig 1. IMUs sensor placement on the dogs’ neck, back, and chest.

https://doi.org/10.1371/journal.pone.0286311.g001

The IMU Raw datasets were concatenated producing 27 data streams, as follows: 3 IMUs (neck, chest, and back) containing 3 sensors each (accelerometer, gyroscope, magnetometer) with 3 axes each (X, Y, Z). The magnetometer features were removed from the dataset as initial experiments confirmed that they were not as predictive of posture. Therefore, the resulting IMU Raw dataset comprised accelerometer and gyroscope measurements which contained a total of 18 columns.

2.4 Data collection protocol

Data collection sessions took place in a room with a test area of 11.5m x 8.5m at the IGDB Training Centre in Cork, Ireland. Firstly, the IMUs were initialised using the local time and placed on the dog to gather inertial data on canine postures. In order to synchronise the IMU data and video recording, the local time was shown at the start of the video recording. Then the behaviour test described in Table 2 was performed to gather information on the key activities.

thumbnail
Table 2. Behaviour Test Protocol for canine posture data monitoring and acquisition.

https://doi.org/10.1371/journal.pone.0286311.t002

Walking, standing, sitting, and lying down were performed following commands, while body shakes occurred spontaneously. Some subjects were more cooperative than others in following the instructions to perform and hold postures for the predetermined period of time. Hence, positive reinforcement methods were administered by the handler using rewards in the form of verbal praise and food to increase adherence to the protocol.

2.5 Data annotation

The postures performed by the dog in the video-recorded session were annotated and classified into two types, and five postures, as described in Table 3.

Posture Timestamps datasets were created for each video-recorded data collection session second by second. They comprised of the timestamp for the start and finish video recording time for each posture described.

The IMU Raw dataset was combined with the Posture Timestamps dataset to form a unified annotated dataset entitled IMU Posture dataset which contains IMU Raw data, dog name and breed, data collection number, and the two labels, namely: posture and type. Transitions and miscellaneous body postures which did not fit any of the previous categories were manually excluded from the dataset. Table 4 shows the number of observations in the IMU Posture dataset per posture and type.

thumbnail
Table 4. Type, posture, and the number of observations in the IMU Posture dataset.

https://doi.org/10.1371/journal.pone.0286311.t004

2.6 Feature extraction

Features were calculated using the IMU Posture dataset considering windows of time where a unique posture was performed. Three parameters control the creation of the window as illustrated in Fig 2 and described below:

  • Transition time (t_time): time between different postures: t_time = 0.25s.
  • Window size (w_size): rolling window size for calculating statistical measures: w_size = 1s.
  • Window offset (w_offset): sampling offsets in the rolling window: w_offset = 0.5s.
thumbnail
Fig 2. Rolling window used for selecting data for feature extraction.

Hyper-parameters shown include re-sampling transition time (t_time) between postures, window size (w_size) used to calculate the statistical measures, and window offset (w_offset) to create separate observations.

https://doi.org/10.1371/journal.pone.0286311.g002

The dataset was created by calculating diverse statistical measures on rolling windows controlled by the re-sampling hyper-parameters described above, following the standard 50% overlap between subsequent frames [9]. The feature set was obtained with tsfel package [49] which calculated 185 different variables for each of the 18 original raw inertial signals in the IMU Posture dataset, resulting in 3,330 features in the final set. These 185 features calculated by tsfel belonged to 60 types, including 26 spectral, 18 temporal, and 16 statistical features.

2.7 Classification models

All classification models were developed using Python (v3.8.1), in particular, tsfel (v0.1.4) was used to extract features, and sklearn (v1.0.2) and imblearn (v0.9.0) libraries were used to implement machine learning techniques. The models were trained on ICHEC’s Kay supercomputer nodes comprising of 2 x 20 core 2.4GHz Intel Xeon Gold 6148 processors [50].

The original dataset was split into three sets named development, golden, and test while controlling for the dogs in order to prevent overlap between datasets. The observation counts for each of these datasets after feature extraction using the rolling window are shown in Table 5. The development set contained 36 dogs and comprised 82% of the observations. The test set contained 5 dogs and consisted of 16% of the observations, and was created prioritising dogs who performed all the postures. The golden dataset contained data from the Golden Retriever dog only and represented the remaining 2% of the observations. The only dataset used for training models was the development set, the golden and test sets were used to analyse the intra-breed and inter-breed performance of the different models.

thumbnail
Table 5. Number of observations after feature extraction per dataset per posture.

https://doi.org/10.1371/journal.pone.0286311.t005

The development set was partitioned to create independent training and validation sets using 10-fold group cross-validation to ensure that those sets did not contain data from the same dogs. The resulting partitioned datasets were used to fit classification models in order to optimise the estimator hyper-parameters.

Select K Best method was utilised to select the best features by calculating ANOVA F-value between the features and labels. Scores given were grouped by sensor position and type to estimate their importance. Random Forest classifier was deployed to fit the development set using all features selected by Select K Best. Feature importance was grouped by the position and type of the sensor, and also by feature domain and type.

A grid search algorithm was utilised to evaluate all combinations of the hyper-parameters shown in Table 6. The grid search algorithm trained estimators individually on the relevant subset of the development set using each of the 1,000 hyper-parameter combinations. In particular, the feature selection algorithm was set to use the top K = 10, 20, 35, 55, and 80 features out of a total of 3,300 initial features extracted by tsfel. It also optimised two hyper-parameters of the Random Forest classifier, namely maximum depth = 3, 5, 7, and 10 and the number of estimators = 25, 50, 100, 250, and 500.

thumbnail
Table 6. Grid search hyper-parameter set for the classifiers.

https://doi.org/10.1371/journal.pone.0286311.t006

The performance metric chosen for optimising the estimators was f1-weighted, as it provides a good evaluation of the model’s ability to predict the correct class in multi-class classification problems with a natural class imbalance. Accordingly, the grid search algorithm chose the hyper-parameter combination used to create the estimators, which yielded the best f1-weighed score when fitted to the development set. Three novel classifiers were built using the optimal estimators selected by the grid search algorithm.

The next subsections outline different architectures used for combining the optimal estimators, in order to predict five postures, including three static postures (lying down, sitting, and standing) and two dynamic activities (walking and body shake). Classifier 1 comprised one estimator to predict the postures directly. Classifier 2 was a cascade classifier composed of 3 estimators, where the first one detected the type of posture (dynamic or static), and the other two further classified the posture. Classifier 3 had two estimators in sequence, where the first one detected body shakes as an anomaly, and the second one classified the remaining 4 postures.

2.7.1 Classifier 1.

Random Forest classifier was fitted to the entire development dataset with all the 5 classes in order to predict posture directly as indicated in Fig 3.

2.7.2 Classifier 2.

This classifier comprised three estimators, namely type, static and dynamic, as shown in Fig 4. Each estimator was trained separately, as outlined below:

  • Type: Random Forest classifier was fitted to the entire development dataset to predict the label ‘type’ as described in Table 3 which comprised of two classes namely, static and dynamic.
  • Static: Random Forest classifier was fitted to a subset of the dataset containing static postures to predict the classes lying down, sitting, and standing.
  • Dynamic: Random Forest classifier was fitted to a subset of the dataset containing dynamic postures to predict the classes walking and body shake.
thumbnail
Fig 4. Type, static and dynamic estimators used in Classifier 2.

https://doi.org/10.1371/journal.pone.0286311.g004

At prediction time, they were arranged sequentially in two stages to estimate the final posture as illustrated in the diagram in Fig 4. The class predicted by the type model in the first stage determined which estimator would be used in the second stage.

2.7.3 Classifier 3.

This classifier comprised two estimators, namely anomaly and normal as indicated in Fig 5. In the first stage, it identified abnormal behaviour (body shake). Other instances went to the second stage where the normal estimator classified all the other four postures in the second stage.

  • Anomaly: Isolation Forest, instead of Random Forest, was fitted to the entire dataset to identify the minority class body shake while all other four postures were assigned to the majority class. The hyper-parameter space searched for comprised of the following: number of estimators = 25, 50, 100, 250, 500, and contamination = 0.005, 0.01, 0.05, 0.1.
  • Normal: Random Forest classifier was fitted to a subset of the dataset containing static postures to predict the classes lying down, sitting, standing, and walking.
thumbnail
Fig 5. Anomaly and normal estimators used in Classifier 3.

https://doi.org/10.1371/journal.pone.0286311.g005

2.8 Model performance

As discussed, a grid search algorithm was used to find the optimal hyper-parameter combination that would yield the model with the highest f1-weighted using 10-fold cross-validation. The optimal hyper-parameter set, and training and validation times were reported for the model that achieved the highest f1-weighted for each of the estimators composing the classifiers. The best estimators were built using the optimal hyper-parameter sets on the entire development set, and then used for predicting the label on unseen data. Intra-breed and inter-breed classification performance was evaluated using the test set (Labrador Retriever and Crosses), and the results are presented in the golden set (Golden Retriever only), respectively.

Besides the optimisation metric f1-weighted, other additional metrics were reported to facilitate the comparison between classifiers presented in this study and classifiers developed in other studies on canine posture recognition. In particular, the following metrics were reported for each of the postures: TPR (True Positive Rate, also known as Sensitivity or Recall), TNR (True Negative Rate, also known as Specificity), Accuracy, PPV (Positive Predictive Value, also known as precision), and f1 score. Hence, the following metrics were reported for each of the classifiers: f1-weighted, f1-macro, and confusion matrix. Confusion matrices were normalised considering the number of true instances in each class, i.e., the total count in the rows of the matrix.

3 Results

Section 3.1 shows the grid search’s chosen best model hyper-parameter combination, cross-validation results on the development set and performance on the test set evaluated individually and independently. Feature contribution was evaluated based on the f-classification score created by Select K Best and Random Forest classifier’s feature importance metric in Classifier 1. Section 3.2 presents the grouped figures as an estimate of the collective contribution of each sensor type and position.

The best models were deployed as indicated in the classifier diagram in Figs 35. They were evaluated on the test set to estimate intra-breed performance in Section 3.3, and the golden set to estimate inter-breed performance in Section 3.4. The performance achieved using the best classifier to predict posture on the combined test and golden sets is shown in Section 3.5.

3.1 Classification models

3.1.1 Classifier 1.

The best hyper-parameters chosen for the Direct estimator were K = 80 for Select K Best, and maximum depth = 10, and number of estimators = 250 for Random Forest. The time taken to build the estimator using the development set with 10-fold cross-validation was (in seconds): training time (M = 340.90, SD = 3.62) and validation time (M = 0.19, SD = 0.01). This model achieved an f1-weighted mean of 0.89 ±0.04 on the validation set, and an f1-weighted 0.89 and f1-macro 0.82 on the test set.

3.1.2 Classifier 2.

The best hyper-parameters chosen for the Type, Dynamic, and Static estimators are shown in Table 7 along with the training and validation times, and the f1-weighted performance metric for the validation and test sets.

thumbnail
Table 7. Classifier 2 model’s hyper-parameter set selected by grid search, the time (in seconds) taken to train and validate the model using the development set, and f1-weighted performance on 10-fold validation sets (mean ± standard deviation) and test set.

https://doi.org/10.1371/journal.pone.0286311.t007

3.1.3 Classifier 3.

The best hyper-parameters chosen by grid search for the Anomaly and Normal estimators are shown in Table 8 along with the training and validation times, and the f1-weighted performance metric for the validation and test set.

thumbnail
Table 8. Classifier 3 model’s hyper-parameter set selected by grid search, and f1-weighted performance on 10-fold validation sets (mean ± standard deviation) and test set.

https://doi.org/10.1371/journal.pone.0286311.t008

3.2 Feature importance

Select K Best used f-classification to calculate the f-score and p-value for each feature taking the posture label into account. The score given was then normalised and grouped by sensor position (back, chest, and neck) and sensor type (accelerometer and gyroscope) to estimate their collective contribution to the model. These results are shown in Table 9 and indicate that the most important sensor position is on the back and the most important sensor type is the accelerometer for distinguishing between different postures.

thumbnail
Table 9. Feature scores calculated by f-classification in Select K Best by IMU position and sensor.

https://doi.org/10.1371/journal.pone.0286311.t009

Table 10 shows the feature importance given by the Random Forest from Classifier 1 on the 80 features chosen by Select K Best. No features derived from the neck sensor were among the 80 highest-ranked ones. Once again, these results indicate that the most important sensor position was on the back and the most important sensor type was the accelerometer.

thumbnail
Table 10. Feature importance calculated by Random Forest classifier considering the 80 features previously selected by Select K Best.

https://doi.org/10.1371/journal.pone.0286311.t010

Table 11 shows more information on the 80 highest-scoring features chosen by Select K best including their domain and collective importance. They belonged to 26 feature types out of the 60 domains calculated, specifically, there were 13 out of 26 spectral, 7 out of 18 temporal, and 5 out of 16 statistical features. In order of importance, the domains providing the best features were the statistical, temporal, and spectral domains. Even though the spectral domain had the biggest number of statistical feature types selected, it provided the least collective importance as calculated by Random Forest.

thumbnail
Table 11. Feature importance calculated by Random Forest classifier considering the 80 features selected by Select K Best by domain.

https://doi.org/10.1371/journal.pone.0286311.t011

3.3 Intra-breed evaluation

A summary of the classification metrics on the test set for Classifiers 1, 2, and 3 per class, and f1-weighted and f1-macro averages is shown in Tables 1214, respectively. The three classifiers achieved very similar results. Classifier 3 achieved the best f1-macro as a result of its improved ability to correctly identify body shakes; however, its f1-weighted was slightly lower for standing and lying down.

thumbnail
Table 12. Classification metrics per posture achieved using the best models selected by grid search in Classifier 1 on the test set.

https://doi.org/10.1371/journal.pone.0286311.t012

thumbnail
Table 13. Classification metrics per posture achieved using the best models selected by grid search in Classifier 2 on the test set.

https://doi.org/10.1371/journal.pone.0286311.t013

thumbnail
Table 14. Classification metrics per posture achieved using the best models selected by grid search in Classifier 3 on the test set.

https://doi.org/10.1371/journal.pone.0286311.t014

The normalised confusion matrices with the predicted labels on the test set using Classifier 1, 2, and 3 models are shown in Figs 68, respectively. The main difference appears in the body shake performance as this class was mainly correctly classified by Classifier 3 model only. This improved performance in the minority class did not negatively affect the model’s ability to identify other classes as they showed comparable figures to Classifiers 1 and 2.

thumbnail
Fig 6. Confusion matrix for Classifier 1 on the test set.

https://doi.org/10.1371/journal.pone.0286311.g006

thumbnail
Fig 7. Confusion matrix for Classifier 2 on the test set.

https://doi.org/10.1371/journal.pone.0286311.g007

thumbnail
Fig 8. Confusion matrix for Classifier 3 on the test set.

https://doi.org/10.1371/journal.pone.0286311.g008

3.4 Inter-breed evaluation

The golden dataset was used to evaluate the performance of each experimental model trained on data from Labrador Retrievers (pure and Golden Retriever crosses) on dogs from another breed (Golden Retriever). A summary of the classification metrics on the golden set for Classifiers 1, 2, and 3 per class, and f1-weighted and f1-macro averages is shown in Tables 1517, respectively. Classifier 3 again achieved the best f1-macro and f1-weighted metrics, it was the only model capable of correctly classifying body shakes.

thumbnail
Table 15. Classification metrics per posture achieved using the best models selected by grid search in Classifier 1 on the golden set.

https://doi.org/10.1371/journal.pone.0286311.t015

thumbnail
Table 16. Classification metrics per posture achieved using the best models selected by grid search in Classifier 2 on the golden set.

https://doi.org/10.1371/journal.pone.0286311.t016

The normalised confusion matrices with the predicted labels for the golden set using Classifier 1, 2, and 3 models are shown in Figs 911, respectively. It can be visually seen that Classifier 3 achieves the overall best results in correctly classifying the classes.

thumbnail
Fig 9. Confusion matrix for Classifier 1 on the golden set.

https://doi.org/10.1371/journal.pone.0286311.g009

thumbnail
Fig 10. Confusion matrix for Classifier 2 on the golden set.

https://doi.org/10.1371/journal.pone.0286311.g010

thumbnail
Fig 11. Confusion matrix for Classifier 3 on the golden set.

https://doi.org/10.1371/journal.pone.0286311.g011

thumbnail
Table 17. Classification metrics per posture achieved using the best models selected by grid search in Classifier 3 on the golden set.

https://doi.org/10.1371/journal.pone.0286311.t017

3.5 Best classifier

Intra-breed and inter-breed evaluation revealed that there were no significant performance differences between Golden Retrievers, Labrador Retrievers, and their crosses. Classifier 3 was selected as the best classifier considering the results from Section 3.1. In order to obtain unified performance metrics for the best model, Classifier 3 was used to calculate performance on the combined test and golden sets.

A summary of the key classification metrics per class, and f1-weighted and f1-macro averages is shown in Table 18; and a confusion matrix is shown in Fig 12. These are considered the final results of the present work.

thumbnail
Table 18. Classification metrics per posture achieved using the best models selected by grid search in Classifier 3 on the test and golden sets combined.

https://doi.org/10.1371/journal.pone.0286311.t018

thumbnail
Fig 12. Confusion matrix for Classifier 3 on the test and golden sets.

https://doi.org/10.1371/journal.pone.0286311.g012

4 Discussion

The advantages of using IMUs over camera systems in the field of behaviour monitoring have already been established [51]. In the present work, a novel posture estimation system was developed specifically for working dogs. Accordingly, the data collection protocol designed includes popular working dog breeds and positions multiple IMUs on the harness. The main advantage of the posture estimation system developed in this work is the superior prediction performance achieved as demonstrated through comparison with previous research. The novel machine learning algorithm developed contributes to the state-of-the-art by employing advanced feature extraction techniques and experimenting with three different cascade architectures including anomaly detection models for the first time.

This was the first time that advanced feature extraction methods were applied to a canine posture recognition system. Feature extraction was performed with the Python package tsfel [49] using a rolling window, resulting in a dataset composed of 3,300 features and 30,224 observations. Features were chosen by the Select K Best algorithm based on the scores given by f-classification considering the label posture. The analysis of the collective contributions of the features deriving from each of the three IMUs placed on the dog’s neck, chest, and back revealed the importance of the back and chest sensors. Moreover, only features from the back and chest IMUs were selected by the algorithm, indicating that these are more informative of the targeted postures than the ones derived from the neck sensor. Brugarolas et al. [5] collected data from IMUs placed in similar positions on 7 dogs; however, only 2 of these dogs had IMUs placed on their chest and back sensors. Even though data were very limited, they suggested the significance of the rump, neck, and chest sensors which was confirmed in this research with a much bigger sample size (N = 42). Accelerometer data were more indicative of posture than gyroscope data as shown in Tables 9 and 10. This finding is in line with previous research [5, 7]. It is important to note that these results are influenced by the types of posture targeted in each study.

A grid search was deployed to find the optimal combination of hyper-parameters for the feature selection and classifier as outlined in Section 2.7. In particular, it used a 10-fold group cross-validation technique to calculate the performance in 10 different subsets of the development set while controlling for dogs to select the hyper-parameters that resulted in the best f1-weighted performance. The main drawback of the grid search algorithm is that it can lead to overfitting as it selects the model that delivers the best mean fold performance and does not take into account the gap between training and test performance or the model complexity.

Three classifier architectures were evaluated in order to improve the overall performance in correctly identifying postures despite the naturally unequal distribution of observations per label. Modifying the classifier architecture resulted in an improvement of the f1-score of the minority class (body shake) while maintaining a comparable performance on the majority classes (lying down, sitting, standing, and walking). This can be particularly observed from Classifier 1 and 3 model results in Tables 12 and 14 on the test set, and Tables 15 and 17 on the golden set. In the latter case, Classifiers 1 and 2 did not appropriately identify the body shake at all, while Classifier 3 correctly labelled all body shakes. In terms of training and validation times, Classifier 1 was the fastest, while Classifiers 2 and 3 took a similar amount of time.

One unexpected result was the superior performance of all classifiers on the golden set as compared to the test set. The only class performing less well was body shake as indicated by the f1-scores in Tables 1214 on the test; and Tables 1517 on the golden set. Interestingly, Gerencser et al. [7] also reported a similar result as the inter-breed model achieved higher performance than the intra-breed, from 70.3% to 73.6%, and 72.6% to 73.5% using only Malinois and Labradors data for training, respectively. This indicates that the variability deriving from the breed difference was less significant than the methodological limitations of the data collection and annotation procedures. Hence, this result suggests that this model can be successfully utilised on not only Labrador Retrievers and crosses with Golden Retrievers but also pure Golden Retrievers as well.

4.1 Comparison with previous work

Classifier 3 was chosen as the best classification model in the present work, and its performance on the test and golden sets combined as shown in Section 3.5 was compared to previous studies in Table 19. The inclusion criteria for such comparison considered only papers that (1) estimated canine posture using IMUs; (2) provided classification metrics per posture (body shake, lying down, sitting, standing, and walking); (3) evaluated performance on different dogs due to significant performance difference between intra-subject and inter-subject metrics [5, 7]. Considering the nine research papers in Table 1, five papers were excluded because they did not meet the criteria [5, 6, 8, 9, 12]. However, studies that utilised commercial systems were included [35, 39, 52].

thumbnail
Table 19. Comparison between Classifier 3 performance and previous studies reporting inter-subject classification performance metrics per posture.

The best geometric means between TPR and TNR (g-mean) are in bold.

https://doi.org/10.1371/journal.pone.0286311.t019

Chambers et al. [39] recommended the use of sensitivity (TPR) and specificity (TNR) to compare the performance of classifiers trained on different datasets in order to control for class imbalance effects. Hence, both metrics were combined in a geometric mean (g-mean) for ease of comparison. The highest g-mean value for each posture was highlighted in bold.

Classifier 3 achieved the highest performance for the postures lying down, sitting, and standing on the combined test and golden sets. The best g-means for body shake and walking were reported by Chambers et al. [39] and Den Uijl et al. [52], respectively. Gerencser et al. [7] achieved the highest TPR for walking; however, there was insufficient information to calculate other performance metrics. It is important to note that, although Ribeiro et al. [11] reported a high TPR for sitting and standing, there were just a few observations from 5 dogs. Moreover, it was not clear whether the algorithm was built using data from the same dogs that were used in the test set.

Classifier 3 achieved the top overall performance on the five postures considering that it attained an arithmetic mean of the f1-score, also known as f1-macro, of 0.83 compared to 0.73 in Chambers et al. [39]. Furthermore, Classifier 3 produced an average TPR of 0.92 compared to 0.905, 0.83, 0.74 in Kumpulainen et al. [38], Gerencser et al. [7], and Ribeiro et al. [11], respectively, taking into account the postures lying down, sitting, standing, and walking. Classifier 3 has outperformed previous models published in the literature for predicting the five postures on dogs that did not belong to the original training set. Conclusively, these results indicate that the proposed system can outperform the state-of-the-art in a real environment.

4.2 Limitations and opportunities

As discussed, inaccurate annotations could account for a significant number of errors. Two main methodological limitations were identified to affect the precision of labels. Firstly, data synchronisation was done by showing the standard time on camera at the beginning of the video recording while this could be improved by shaking the sensor in front of the camera. Secondly, data annotation was done second by second, while it could be improved by using some video annotation software to increase the precision of the timestamps. These limitations can result in labels being misaligned with the data.

Other two causes of incorrect labels were identified: undefined transitions between two activities and rapid changes from one activity to another. The former case refers to the variability in the duration of transitions [11], for instance transitioning from walking to standing is nearly instantaneous; however, transitioning from standing to lying down can take a couple of seconds; such an observation could understandably be classified as sitting. The latter is observed in cases where the dog is walking but stands briefly only to resume walking again, such an observation could acceptably be classified as standing. This could explain the misclassification between such classes.

In order to address the first issue, a fixed transition time between postures was used when extracting features; however, it could still not have been long enough for short postures. The second issue is harder to address as sometimes the posture itself may not be very well defined. It is advisable to utilise video commenting software for data labelling, allowing higher precision in the annotations.

The minority class body shake was sometimes incorrectly classified as walking. Such results can be attributed to two main causes. Firstly, it is reasonable to estimate that a significant number of miss-classification cases could have resulted from inaccurate labels. This is because body shakes are very rapid movements typically happening between walking or standing postures. As annotations were made second by second, it is possible that some frames contain mixed postures. Secondly, these are the only dynamic postures while all others are static. It can be assumed that there exists a higher level of similarity between these postures when compared to all other postures, making it harder to distinguish them.

Different window sizes could be employed as some classes like body shake are shorter in duration than others. Accordingly, other techniques could be explored including KNN with Dynamic Time Warping [12], and deep learning methods such as convolutional neural network (CNN) [53] and long short-term memory (LSTM) motifs [39].

4.3 Recommendations

We suggest that researchers use the framework employed in this research. In particular, they are requested to make their datasets publicly available, describe key methodological details and expand their performance reports. This will allow for easier comparison, reproducibility, and knowledge transfer for the advancement of this field.

The following details are crucial and should be reported: data collection materials (sensor type and sampling rate); methods (number of dogs, breeds, context, types of behaviours, duration, repetitions); data pre-processing (window size and overlap); feature extraction (calculation, hyper-parameters); dataset statistics (size of training, development, and validation/test sets); dataset splitting criteria (e.g. breed, subject, number of observations); dataset details (number of observations per behaviour); prediction algorithm (type, hyper-parameters); optimisation (hyper-parameters and search space); and evaluation (technique, hyper-parameters, metrics, training, and testing time).

Future research should also test the performance of new dog activity recognition models on previously published datasets whenever possible to benchmark results. Researchers are encouraged to validate commercial canine activity monitoring devices, in terms of their performance in quantifying activity levels and identifying typical behaviours.

5 Conclusions

The main goal of this study was to investigate whether the present system comprising neck, back, and chest sensors would surpass the performance reported in published work on canine posture classification systems, namely: body shake, lying down, sitting, standing, and walking. This work contributes to the state-of-the-art knowledge in using IMUs for posture prediction using machine learning, by expanding the current understanding of sensor position, feature extraction, and importance as well as model architecture.

The importance of adding IMU sensors to the back and chest for more accurate posture prediction was demonstrated, encouraging the inclusion of IMU sensors on the dog harness for applications where the accuracy of posture predictions is critical, as in the case of some types of working dogs. Advanced feature extraction techniques for time series were been successfully employed and validated for the first time in the field of canine posture prediction. Feature selection was necessary to remove uninformative features, reducing the complexity of models and processing time.

The best-performing model architecture was Classifier 3 which comprised an anomaly detection estimator to identify the minority class (body shake) followed by a classifier to detect the other four postures (lying down, sitting, standing, and walking). The noticeably higher performance achieved by this novel classifier proves the advantage of combining different classifiers to leverage their respective strengths. In particular, this result indicates that the cascade machine learning architecture including the anomaly detection model significantly improved the detection rate of less frequent behaviour (body shake).

The novel canine posture estimator presented here achieved the best overall performance for predicting the five postures compared to previous research. The best performing canine posture estimator previously reported in the literature was presented in Chambers et al. [39]. This novel estimator attained a superior performance in terms of both f1-macro—0.83 versus 0.73—and mean g-mean—0.91 versus 0.84, respectively. Hence, this novel posture estimator system for working dogs provides the most correct predictions of five canine behaviours.

Finally, the IMU Posture dataset built using 42 working dogs was made publicly available. This allows the application of other machine learning techniques and a fair performance comparison between different methods. In particular, future research should investigate the use of other advanced feature selection techniques [54, 55], improve the explainability of models using techniques such as local interpretable model agnostic (LIME) and ELI5 [56, 57], and deep learning techniques (CNNs and LSTM networks) for the prediction of postures. Additional data should be added to the dataset to include different postures and other working dog breeds to create more complete automatic ethograms. These would enable the development of a range of applications to assist working dog organisations during dogs’ training and working lives.

Acknowledgments

The authors wish to acknowledge the support received from the Irish Guide Dogs for the Blind, the computational resources provided by the Irish Centre for High-End Computing (ICHEC), and data collection support from students at UCC, especially MariaVittoria Lorenzi, Hazel Craven, Conghal Hewson, and Jennifer Kenny.

References

  1. 1. Jones AC, Gosling SD. Temperament and Personality in Dogs (Canis Familiaris): A Review and Evaluation of Past Research. Applied Animal Behaviour Science. 2005;95(1-2):1–53.
  2. 2. Gartner MC. Pet Personality: A Review. Personality and Individual Differences. 2015;75:102–113.
  3. 3. Bodnariu A. Indicators of Stress and Stress Assessment in Dogs. Lucrari Stiinfice: Medicina Veterinara Timisoara (Scientifical Papers: Veterinary Medicine). 2008.
  4. 4. Bozkurt A, Roberts DL, Sherman BL, Brugarolas R, Mealin S, Majikes J, et al. Toward Cyber-Enhanced Working Dogs for Search and Rescue. IEEE Intelligent Systems. 2014;29(6):32–39.
  5. 5. Brugarolas R, Loftin RT, Yang P, Roberts DL, Sherman B, Bozkurt A. Behavior Recognition Based on Machine Learning Algorithms for a Wireless Canine Machine Interface. In: IEEE International Conference on Body Sensor Networks (BSN). Cambridge, MA, USA: IEEE; 2013.
  6. 6. Brugarolas R, Roberts D, Sherman B, Bozkurt A. Machine Learning Based Posture Estimation for a Wireless Canine Machine Interface. In: 2013 IEEE Topical Conference on Biomedical Wireless Technologies, Networks, and Sensing Systems. IEEE; 2013. p. 10–12.
  7. 7. Gerencsér L, Vásárhelyi G, Nagy M, Vicsek T, Miklósi A. Identification of Behaviour in Freely Moving Dogs (Canis Familiaris) Using Inertial Sensors. PLoS ONE. 2013;8(10):e77814. pmid:24250745
  8. 8. Kumpulainen P, Gizatdinova Y, Vehkaoja A, Valldeoriola A, Somppi S, Törnqvist H, et al. Dog Activity Classification with Movement Sensor Placed on the Collar. In: Proceedings of the Fifth International Conference on Animal-Computer Interaction—ACI’18. Atlanta, Georgia: ACM Press; 2018. p. 1–6.
  9. 9. Ladha C, Hammerla N, Hughes E, Olivier P, Ploetz T. Dog’s Life: Wearable Activity Recognition for Dogs. In: Proceedings of the 2013 ACM International Joint Conference on Pervasive and Ubiquitous Computing—UbiComp’13. ACM Press; 2013. p. 415.
  10. 10. Majikes J, Brugarolas R, Winters M, Yuschak S, Mealin S. Balancing Noise Sensitivity, Response Latency, and Posture Accuracy for a Computer-Assisted Canine Posture Training System. International Journal of Human-Computer Studies. 2017.
  11. 11. Ribeiro C, Ferworn A, Denko M, Tran J. Canine Pose Estimation: A Computing for Public Safety Solution. In: 2009 Canadian Conference on Computer and Robot Vision. IEEE; 2009. p. 37–44.
  12. 12. Kiyohara T, Orihara R, Sei Y, Tahara Y, Ohsuga A. Activity Recognition for Dogs Using Off-the-Shelf Accelerometer:. In: Proceedings of the International Conference on Agents and Artificial Intelligence. Lisbon, Portugal: SCITEPRESS—Science and and Technology Publications; 2015. p. 100–110.
  13. 13. Pons P, Jaen J, Catala A. Detecting Animals’ Body Postures Using Depth-Based Tracking Systems. In: Proceedings of Measuring Behavior. Dublin, Ireland; 2016. p. 5.
  14. 14. Mealin S, Domínguez IX, Roberts DL. Semi-Supervised Classification of Static Canine Postures Using the Microsoft Kinect. In: Proceedings of the Third International Conference on Animal-Computer Interaction. Milton Keynes United Kingdom: ACM; 2016. p. 1–4.
  15. 15. Klishkovskaia T, Aksenov A, Sinitca A, Zamansky A, Markelov OA, Kaplun D. Development of Classification Algorithms for the Detection of Postures Using Non-Marker-Based Motion Capture Systems. Applied Sciences. 2020;10(11):4028.
  16. 16. Tsai MF, Huang JY. Predicting Canine Posture With Smart Camera Networks Powered by the Artificial Intelligence of Things. IEEE Access. 2020;8:220848–220857.
  17. 17. Arora T, Grey I. Health Behaviour Changes during COVID-19 and the Potential Consequences: A Mini-Review. Journal of Health Psychology. 2020;25(9):1155–1163. pmid:32551944
  18. 18. Morgan L, Protopopova A, Birkler RID, Itin-Shwartz B, Sutton GA, Gamliel A, et al. Human–Dog Relationships during the COVID-19 Pandemic: Booming Dog Adoption during Social Isolation. Humanities and Social Sciences Communications. 2020;7(1):155.
  19. 19. Väätäjä H, Majaranta P, Cardó AV, Isokoski P, Somppi S, Vehkaoja A, et al. The Interplay Between Affect, Dog’s Physical Activity and Dog–Owner Relationship. Frontiers in Veterinary Science. 2021;8:673407. pmid:34957271
  20. 20. Jones S, Dowling-Guyer S, Patronek GJ, Marder AR, Segurson D’Arpino S, McCobb E. Use of Accelerometers to Measure Stress Levels in Shelter Dogs. Journal of Applied Animal Welfare Science. 2014;17(1):18–28. pmid:24484308
  21. 21. Belda B, Enomoto M, Case BC, Lascelles BDX. Initial Evaluation of PetPace Activity Monitor. The Veterinary Journal. 2018;237:63–68. pmid:30089547
  22. 22. Ribeiro C, Ferworn A, Denko M, Tran J, Mawson C. Wireless Estimation of Canine Pose for Search and Rescue. In: 2008 IEEE International Conference on System of Systems Engineering. IEEE; 2008. p. 1–6.
  23. 23. Britt WR, Miller J, Waggoner P, Bevly DM, Hamilton JA. An Embedded System for Real-Time Navigation and Remote Command of a Trained Canine. Personal and Ubiquitous Computing. 2011;15(1):61–74.
  24. 24. Mealin S, Foster M, Cleghern Z, Bozkurt A, Roberts DL. Using Inertial Measurement Unit Data for Objective Evaluations of Potential Guide Dogs. In: Proceedings of the Seventh International Conference on Animal-Computer Interaction. Milton Keynes United Kingdom: ACM; 2020. p. 1–11.
  25. 25. Brugarolas R, Roberts D, Sherman B, Bozkurt A. Posture Estimation for a Canine Machine Interface Based Training System. In: 34th Annual International Conference of the IEEE EMBS. IEEE; 2012. p. 4489–4492.
  26. 26. Majikes JJ, Yuschak S, Walker K, Brugarolas R, Mealin S, Foster M, et al. Stimulus Control for Semi-autonomous Computer Canine-Training. In: Mangan M, Cutkosky M, Mura A, Verschure PFMJ, Prescott T, Lepora N, editors. Biomimetic and Biohybrid Systems. vol. 10384. Cham: Springer International Publishing; 2017. p. 279–290.
  27. 27. Barthélémy I, Barrey E, Thibaud JL, Uriarte A, Voit T, Blot S, et al. Gait Analysis Using Accelerometry in Dystrophin-Deficient Dogs. Neuromuscular Disorders. 2009;19(11):788–796. pmid:19800232
  28. 28. Hansen BD, Lascelles BDX, Keene BW, Adams AK, Thomson AE. Evaluation of an Accelerometer for At-Home Monitoring of Spontaneous Activity in Dogs. American Journal of Veterinary Research. 2007;68(5):468–475. pmid:17472445
  29. 29. Brown DC, Boston RC, Farrar JT. Use of an Activity Monitor to Detect Response to Treatment in Dogs with Osteoarthritis. Journal of the American Veterinary Medical Association. 2010;237(1):66–70. pmid:20590496
  30. 30. Nuttall T, McEwan N. Objective Measurement of Pruritus in Dogs: A Preliminary Study Using Activity Monitors. Veterinary Dermatology. 2006;17(5):348–351. pmid:16961821
  31. 31. Muñana KR, Nettifee JA, Griffith EH, Early PJ, Yoder NC. Evaluation of a Collar-mounted Accelerometer for Detecting Seizure Activity in Dogs. Journal of Veterinary Internal Medicine. 2020;34(3):1239–1247. pmid:32293068
  32. 32. Beerda B, Schilder MBH, van Hooff JARAM, de Vries HW. Manifestations of Chronic and Acute Stress in Dogs. Applied Animal Behaviour Science. 1997;52(3-4):307–319.
  33. 33. Part CE, Kiddie JL, Hayes WA, Mills DS, Neville RF, Morton DB, et al. Physiological, Physical and Behavioural Changes in Dogs (Canis Familiaris) When Kennelled: Testing the Validity of Stress Parameters. Physiology & Behavior. 2014;133:260–271. pmid:24866912
  34. 34. Tomkins LM, Thomson PC, McGreevy PD. Behavioral and Physiological Predictors of Guide Dog Success. Journal of Veterinary Behavior: Clinical Applications and Research. 2011;6(3):178–187.
  35. 35. den Uijl I, Gómez Álvarez CB, Bartram D, Dror Y, Holland R, Cook A. External Validation of a Collar-Mounted Triaxial Accelerometer for Second-by-Second Monitoring of Eight Behavioural States in Dogs. PLOS ONE. 2017;12(11):e0188481. pmid:29186154
  36. 36. Mealin S, Howell S, Roberts DL. Towards Unsupervised Canine Posture Classification via Depth Shadow Detection and Infrared Reconstruction for Improved Image Segmentation Accuracy. In: Lepora NF, Mura A, Mangan M, Verschure PFMJ, Desmulliez M, Prescott TJ, editors. Biomimetic and Biohybrid Systems. Lecture Notes in Computer Science. Cham: Springer International Publishing; 2016. p. 155–166.
  37. 37. Ferdinandy B, Gerencsér L, Corrieri L, Perez P, Újváry D, Csizmadia G, et al. Challenges of Machine Learning Model Validation Using Correlated Behaviour Data: Evaluation of Cross-Validation Strategies and Accuracy Measures. PLOS ONE. 2020;15(7):e0236092. pmid:32687528
  38. 38. Kumpulainen P, Cardó AV, Somppi S, Törnqvist H, Väätäjä H, Majaranta P, et al. Dog Behaviour Classification with Movement Sensors Placed on the Harness and the Collar. Applied Animal Behaviour Science. 2021;241:105393.
  39. 39. Chambers RD, Yoder NC, Carson AB, Junge C, Allen DE, Prescott LM, et al. Deep Learning Classification of Canine Behavior Using a Single Collar-Mounted Accelerometer: Real-World Validation. Animals. 2021;11(6):1549. pmid:34070579
  40. 40. Marcato M, Kenny J, O’Riordan R, O’Mahony C, O’Flynn B, Galvin P. Assistance Dog Selection and Performance Assessment Methods Using Behavioural and Physiological Tools and Devices. Applied Animal Behaviour Science. 2022;254:105691.
  41. 41. Chen Y, Hu B, Keogh E, Batista GEAPA. DTW-D: Time Series Semi-Supervised Learning from a Single Example. In: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Chicago Illinois USA: ACM; 2013. p. 383–391.
  42. 42. Beerda B, Schilder MBH, van Hooff AM, de Vries HW. Behavioural and Hormonal Indicators of Enduring Environmental Stress in Dogs. Animal Welfare. 2000; p. 15.
  43. 43. Ennik I, Liinamo AE, Leighton E, van Arendonk J. Suitability for Field Service in 4 Breeds of Guide Dogs. Journal of Veterinary Behavior. 2006;1(2):67–74.
  44. 44. Evans KM, Lewis TW, Asher L, Blythe S, Bottomley M, Tootill L, et al. Genetic Evaluation of Traits in a Standardized Behavioral Test for Potential Guide Dog Puppies Using Crossbreed Models. Journal of Veterinary Behavior. 2015;10(6):459–464.
  45. 45. van Rooy D, Wade CM. Behavioral Genetics of Dog Breeds. In: Choe JC, editor. Encyclopedia of Animal Behavior (Second Edition). Oxford: Academic Press; 2019. p. 312–322.
  46. 46. Wilkinson MD, Dumontier M, Aalbersberg IJ, Appleton G, Axton M, Baak A, et al. The FAIR Guiding Principles for Scientific Data Management and Stewardship. Scientific Data. 2016;3(1):160018. pmid:26978244
  47. 47. Vehkaoja A, Somppi S, Törnqvist H, Valldeoriola Cardó A, Kumpulainen P, Väätäjä H, et al. Description of Movement Sensor Dataset for Dog Behavior Classification. Data in Brief. 2022;40:107822. pmid:35079615
  48. 48. Marcato M, Tedesco S, O’Mahony C, O’Flynn B, Galvin P. Inertial sensor dataset for Dog Posture Recognition. Mendeley Data, V1.
  49. 49. Barandas M, Folgado D, Fernandes L, Santos S, Abreu M, Bota P, et al. TSFEL: Time Series Feature Extraction Library. SoftwareX. 2020;11:100456.
  50. 50. ICHEC. Kay; 2022. https://www.ichec.ie/about/infrastructure/kay.
  51. 51. Stiles EK, Palestrini C, Beauchamp G, Frank D. Physiological and Behavioral Effects of Dextroamphetamine on Beagle Dogs. Journal of Veterinary Behavior. 2011;6(6):328–336.
  52. 52. den Uijl I, Gomez Alvarez C, Dror I, Manning N, Bartram D, Cook A. Validation of a Collar-Mounted Accelerometer That Identifies Eight Canine Behavioural States, Including Those with Dermatologic Significance. In: Validation of a Collar-Mounted Accelerometer That Identifies Eight Canine Behavioural States, Including Those with Dermatologic Significance. Weybridge, UK; 2016. p. 81–84.
  53. 53. Sahoo JP, Prakash AJ, Pławiak P, Samantray S. Real-Time Hand Gesture Recognition Using Fine-Tuned Convolutional Neural Network. Sensors. 2022;22(3):706. pmid:35161453
  54. 54. Patro KK, Jaya Prakash A, Jayamanmadha Rao M, Rajesh Kumar P. An Efficient Optimized Feature Selection with Machine Learning Approach for ECG Biometric Recognition. IETE Journal of Research. 2022;68(4):2743–2754.
  55. 55. Sinha VK, Patro KK, Pławiak P, Prakash AJ. Smartphone-Based Human Sitting Behaviors Recognition Using Inertial Sensor. Sensors. 2021;21(19):6652. pmid:34640971
  56. 56. Javed AR, Ahmed W, Pandya S, Maddikunta PKR, Alazab M, Gadekallu TR. A Survey of Explainable Artificial Intelligence for Smart Cities. Electronics. 2023;12(4):1020.
  57. 57. Javed AR, Khan HU, Alomari MKB, Sarwar MU, Asim M, Almadhor AS, et al. Toward Explainable AI-empowered Cognitive Health Assessment. Frontiers in Public Health. 2023;11:1024195. pmid:36969684