skip to main content
research-article
Open Access

Multi-modal Open World User Identification

Authors Info & Claims
Published:18 October 2021Publication History

Skip Abstract Section

Abstract

User identification is an essential step in creating a personalised long-term interaction with robots. This requires learning the users continuously and incrementally, possibly starting from a state without any known user. In this article, we describe a multi-modal incremental Bayesian network with online learning, which is the first method that can be applied in such scenarios. Face recognition is used as the primary biometric, and it is combined with ancillary information, such as gender, age, height, and time of interaction to improve the recognition. The Multi-modal Long-term User Recognition Dataset is generated to simulate various human-robot interaction (HRI) scenarios and evaluate our approach in comparison to face recognition, soft biometrics, and a state-of-the-art open world recognition method (Extreme Value Machine). The results show that the proposed methods significantly outperform the baselines, with an increase in the identification rate up to 47.9% in open-set and closed-set scenarios, and a significant decrease in long-term recognition performance loss. The proposed models generalise well to new users, provide stability, improve over time, and decrease the bias of face recognition. The models were applied in HRI studies for user recognition, personalised rehabilitation, and customer-oriented service, which showed that they are suitable for long-term HRI in the real world.

Skip 1INTRODUCTION Section

1 INTRODUCTION

User identification is an important step towards achieving and maintaining a personalised long-term interaction with robots. For instance, a user may need to be identified for providing personalised rehabilitation therapy [41]. When a robot is first deployed, it will start from a “tabula rasa” state with no prior knowledge of users. As users are encountered over a possibly extended period of time, their identity and information are stored by the robot. Hence, the system has to identify enrolled and “unknown” users, which is known as open-set identification. Open-set identification is a well-established field [48, 76, 77], but in a real-world setting, these unknown users might need to be added into the system for future recognition. One solution is to retrain the system after introducing a novel user. However, this requires storing the previous samples, which could create a prohibitively large computational burden in long-term deployments. Furthermore, it would require a significant amount of time to retrain with a growing number of users and samples [8]. Instead, the system should allow scaling and support incremental learning of new classes, which is termed open world recognition [8].

Face recognition (FR), i.e., identifying a person based on their face, has been the most prominent technique in biometric identification due to its non-intrusive character. Most state-of-the-art methods use deep learning-based approaches [68, 79, 80, 81], but only a few approaches exist for open-set recognition [9, 33]. Most models are not suitable for open world recognition due to the catastrophic forgetting problem, which refers to the drastic loss of performance on previously learned classes when a new class is introduced [62, 63, 66]. Existing approaches that could help to overcome this problem often require a part of the previous data for retraining, which might not be available.

Incremental learning is not sufficient for adapting to changes in the environment. For instance, an algorithm designed for open world recognition may not be able to recognise a person after a new haircut, because the model is not updated for known samples. Humans show a good model for recognition, because they can continuously adapt to changing circumstances by updating their prior beliefs, known as online learning, and use multi-modal information instead of a single biometric (modality) for estimation of the identity, such as recognising a person from the voice in a dark room. Biometric systems that combine multiple biometric traits or attributes obtained through the same sensor (e.g., face and iris [16, 21, 83, 87]) or various sensors (e.g., face and voice [10, 17, 18, 58, 82]) for establishing identity are known as multi-modal biometric systems [24, 47]. Most robots are also suitable for multi-modal recognition, as they have multiple sensors and perception algorithms (as shown in Figure 1), which allow them to recognise users even when data are inaccurate or noisy, for example, in the case of image blur or illumination changes [85]. Moreover, the combination of multi-modal data can help overcome issues related to similarities between users1 by differentiating on additional available information, for example, age and gender. Such ancillary physical or behavioural characteristics, called soft biometrics, can be used to improve the recognition performance [24, 45, 47]. Combining multi-modal recognition with online learning can improve recognition further in time. For instance, a user can be initially mistaken for another in certain circumstances, but these variations can be learned over time and combined with other modalities to improve recognition where FR fails.

Fig. 1.

Fig. 1. Robots can make use of multi-modal information to recognise users more accurately in long-term interactions.

In our earlier work [43], we proposed a multi-modal weighted Bayesian Network with online learning, which is the first approach for combining soft biometrics (gender, age, height, and time of interaction) with a primary biometric (face recognition) for open world user identification in real-time human-robot interaction (HRI). This model, here referred to as Multi-modal Incremental Bayesian Network (MMIBN), is the first method for sequential and incremental learning in open world user recognition that allows starting from a state without any known users (i.e., it does not require preliminary training to recognise users and it can learn new users incrementally). This work showed that the proposed model is suitable for real-world human-robot interaction experiments for user recognition in real-time. However, the limited population size (14 users) and the narrow age range (24–40) of the users in that experiment prevented us from claiming that the results can be generalised for application in larger populations. However, obtaining a dataset that encapsulates a diverse set of characteristics for a large number of users over long-term interactions is a laborious task in HRI. Thus, we created the Multi-modal Long-Term User Recognition Dataset,2 which contains images of 200 users (with age range 10 to 63) with name, gender, age, and height labels, along with artificially generated height estimations and various time of interactions to simulate a long-term HRI scenario. We obtained the images from the largest publicly available dataset of face images with gender and age labels,

IMDB-WIKI dataset3 [71, 72]. To obtain the multi-modal biometric information from these images (face, gender, and age estimations), we used (NAOqi) proprietary algorithms of the Pepper robot,4 similar to our earlier work.

Our main contribution is the extension of our earlier work [43] to take in multi-modal information, typically available in HRI, to markedly increase user identification and subsequently improve user experience in long-term interactions for a large number of users in a variety of settings. We also provide a detailed description of the Multi-modal Incremental Bayesian Network, highlighting the mathematical formulations and assumptions behind the models that were not addressed in Reference [43]. In addition, we present our findings from applying the optimised models in long-term HRI experiments in the real world [41, 42, 43]. Correspondingly, we make the following contributions (source code, multi-modal dataset, trained models, and results on the dataset are available online\( ^{2} \)):

  • creating the Multi-modal Long-term User Recognition Dataset with 200 users of varying characteristics;

  • introducing long-term recognition performance loss;

  • combining optimal normalisation methods for each parameter in the Bayesian network in a hybrid approach;

  • formulating the proposed online learning in terms of Expectation Maximization (EM) and Maximum Likelihood (ML);

  • applying Bayesian optimisation on the weights of the soft biometric identifiers and the quality of the estimation;

  • evaluating the proposed model against a state-of-the-art open world recognition method (Extreme Value Machine [73]);

  • evaluating the stability of the model for learning users sequentially (similar to batch learning) and at random intervals (similar to a real-world scenario);

  • evaluating the generalisability of the model for new users (performance during training set in comparison to open-set and closed-set recognition);

  • evaluating the model for varying frequency of user appearances (modelled with uniform and Gaussian timing of interaction, and varying dataset sizes);

  • evaluating the progress of the model over time (with the increasing number of recognitions);

  • analysing recognition bias in face recognition, the proposed approach, and Extreme Value Machine;

  • evaluating the models on the data from the real-world HRI study (four weeks) in Reference [43] in comparison to the corresponding optimised models;

  • evaluating the model in a real-world (five-day) HRI study with a personalised barista robot at an international student campus in Paris (France);

  • evaluating the models in a long-term (five months) HRI study within a cardiac rehabilitation programme at a hospital in Bogotá (Colombia).

The rest of the article is organised as follows: Section 2 gives a brief overview of the current practice of open world recognition, online learning, multi-modal biometrics algorithms, and user recognition in HRI. Section 3 describes the methodology and the structure of the proposed Bayesian network. Section 4 describes the recognition module for NAOqi that is used to obtain the multi-modal biometric information for the proposed model.

Section 5 explains the procedure of the creation of the Multi-modal Long-term User Recognition Dataset. Section 6 presents the empirical evaluation of the proposed methods on closed-set and open-set datasets. Section 7 highlights the implications of the results and discusses the initial assumptions. Section 8 evaluates the optimised models in long-term HRI studies in the real world. Section 9 concludes with a summary of the work.

Skip 2RELATED WORK Section

2 RELATED WORK

Our work lies at the intersection of open world recognition, online learning, multi-modal biometrics, and HRI.

2.1 Open World Recognition

One of the first algorithms applied to open world recognition was Nearest-Non Outlier (NNO) [8], which modified Nearest Class Mean (NCM) [64] for open-set classification and incremental learning. Another approach is Extreme Value Machine (EVM) [73] based on Extreme Value Theory, which outperformed NNO on the open world ImageNet benchmark [8, 73]. However, both of these methods work with incrementally adding a batch of new classes (e.g., 100 at a time), as opposed to incremental learning of classes (one at a time). Similarly, the approach proposed in Reference [29] is based on a center-based similarity space learning method and 1-vs-rest strategy of Support Vector Machines (SVM) for object classification. However, none of these methods has been evaluated on user recognition.

2.2 Online Learning

Several online learning methods exist for various application areas [34]. In video-based recognition, Lee and Kriegman [55] proposed an online learning algorithm of probabilistic appearances, but a prior generic model is necessary for this approach. Boucenna et al. [13] used online and incremental learning in two neural networks for facial expression recognition and face/non-face discrimination in an HRI imitation game. The former neural network uses a k-means variant SAW (Self Adaptive Winner takes all) [49] to categorise focus points in the image, whereas the latter predicts the interaction rhythm [3] (i.e., timing for interaction) to detect whether the user is interacting with the robot. While the face discrimination method was shown to generalise to new users successfully, the facial expression recognition achieved low success rates for generalisation. In addition, both approaches required preliminary training and were evaluated on a low number of users (20). De Rosa et al. [26] used online learning in open world (object) recognition for incremental learning of a classification metric, the threshold for novelty detection, and describing the space of classes. The approach was applied to three existing algorithms, namely, NCM, NNO, and Nearest Ball Classifier (NBC) [27]. Their results showed that online learning increases classification performance.

2.3 Multi-modal Biometrics

In a multi-modal biometric system, information from different identifiers, such as face recognition or gender identification, is fused via prior or post classification [44]. Prior classification requires access to the features or sensor values of the identifiers, which are generally not available for proprietary algorithms.

For post-classification, two approaches exist: classification and combination of confidence scores. Classification methods, such as neural networks and SVM, combine non-homogeneous data from individual classifiers into a feature vector for further classification without the need for preprocessing. In the combination approach, individual matching scores from the identifiers are combined into a scalar score in three steps: (1) normalisation of scores into a common domain, (2) combination of scores based on Bayes decision rule and posterior probabilities, e.g., sum or product rule, and (3) thresholding for classification. The performance of these approaches depends on the chosen method and threshold.

Bayesian approaches have been widely used for combining primary biometrics, such as face and speaker recognition [10, 17, 82], as well as combining soft biometrics [25, 45, 46, 67, 78, 86]. For instance, Jain et al. [45] proposed a Bayesian network for combining fingerprints with soft biometric traits, namely, gender, ethnicity, and height. They used a fixed weighting scheme, where the biometrics with smaller variability and more substantial distinguishing capability were given more weight and achieved slight improvement in recognition. Similarly, Scheirer et al. [78] used a Bayesian network with Noisy-OR weighting that combines face recognition with ethnicity, hair colour, gender, age, eyebrow type, and non-soft biometric contextual information, such as the occupation and location of the person. Contrary to the work in Reference [45] and our approach, they used the accuracy of estimators to adjust the FR match score.

2.4 User Recognition in Human-Robot Interaction

Similar to biometric recognition, the most common approach for user recognition in HRI is through FR [4, 5, 23, 32, 38]. However, robots can take advantage of multi-modal recognition due to the variety of different sensors they carry. Soft biometrics are especially important, because they allow non-intrusive recognition, but only a few studies use soft biometrics. Martinson et al. [61] used a weighted summation of soft biometrics (clothing, complexion, and height) to identify users within a short-term interaction from a group of only three users. Boucenna et al. [12] gathered extensive data (100 images per person) during a game and later evaluated the recognition offline using a Hebbian rule-based neural network. Ouellet et al. [65] combined face recognition, speaker identification, and human metrology through Hampel estimators in closed-set identification using a substantial time for training (3.5 minutes) and a small number of participants (pretraining on 22, test on 7). Al-Qaderi and Rad [1] combined face, body, and speech information using a spiking neural network in closed-set identification and have evaluated on a simulated dataset. These approaches do not apply to open world recognition, hence, their methods are not easily comparable to ours.

Our previous work [43] introduced a multi-modal weighted Bayesian network, which is the first approach in combining multi-modal biometric information for sequential and incremental learning of new users for open world recognition that allows starting from a state without any known users. It is also the first approach in combining soft biometrics (gender, age, height, and time of interaction) with a primary biometric (FR) to identify a user in real-time HRI. Online learning was used for learning the likelihoods of the network from sequential data to improve the recognition over long-term interactions. The weights of the network were optimised to minimise the number of incorrect recognitions. The quality of the estimation measure was introduced to decrease the number of incorrect recognitions for unknown users. The results obtained in a user study with 14 participants over four weeks showed a slight improvement in identification rate (up to 1.4% in open-set and 4.4% in closed-set recognition) compared to 90.3% of FR. The optimised weights suggested that age is the least effective soft biometric parameter, whereas height is the most effective one. Moreover, the Bayesian network performed worse with online learning. However, we concluded that the dataset might be biased towards the participants’ characteristics due to the low number of participants and limited age range, and an evaluation with a bigger dataset is necessary to understand the capabilities of the system entirely.

This article extends the work in Reference [43] for evaluating the approach within the Multi-modal Long-term User Recognition Dataset and two other real-world HRI experiments, and optimising the weights of the Bayesian network through a long-term recognition performance loss criterion with hybrid normalisation.

Skip 3MULTI-MODAL INCREMENTAL BAYESIAN NETWORK Section

3 MULTI-MODAL INCREMENTAL BAYESIAN NETWORK

A Bayesian network is a probabilistic graphical model that represents conditional dependencies of a set of variables through a directed acyclic graph. Bayesian networks are suitable for combining scores of identifiers with uncertainties when the knowledge of the world is incomplete [78].

We developed a weighted multi-modal incremental Bayesian network (MMIBN), integrating multi-modal biometric information for reliable recognition in open world identification through a naive Bayes model (see Figure 2). The naive Bayes classifier model assumes conditional independence between predictors, which is a reasonable assumption for a multi-modal biometric identifier, as the individual identifiers do not affect each other’s results.

Fig. 2.

Fig. 2. The naive Bayesian network model with identity ( \( I \) ), face ( \( F \) ), gender ( \( G \) ), age ( \( A \) ), height ( \( H \) ), and time of interaction ( \( T \) ) nodes.

The architecture for the estimation of the user identity (\( I \)) in MMIBN and the recognition process are presented in Figures 14 and 15 in Appendix A. The primary biometric in our system is face recognition (\( F \)), which is fused with soft biometrics, namely, gender (\( G \)), age (\( A \)), and height (\( H \)) estimations, in addition to the time of interaction (\( T \)), which can be distinguishing if the users are encountered at patterned interaction times, such as for weekly appointments in rehabilitation.

We hypothesise that the integration of these soft biometrics will reduce the effects of noisy data, as described in Section 1, and increase the identification rate. Nonetheless, the MMIBN allows extension with other primary biometric traits, such as voice and fingerprint, and other soft biometrics, such as eye colour and gait, to improve recognition.

The pyAgrum5 [36] library is used for implementing the Bayesian network structure. Parts of MMIBN were previously described in our prior work [43], however, this section provides the underlying mathematical formulations and full details of the system for reproducibility, and introduces the long-term recognition performance loss (Section 3.6) and hybrid normalisation (Section 3.7).

3.1 Structure

The number of states for each node depends on the modality: F and \( I \) nodes have \( n_e \)+1 states, where \( n_e \) is the number of enrolled (known) users. \( A \) and \( H \) nodes are restricted to the available range of the identifier, such as \( [0,75] \) for \( A \) and \( [50,240] \) (cm) for \( H \). \( G \) has “female” and “male” states. \( T \) is defined by the day of the week and the time, through time slots. For example, if each minute corresponds to a time slot (i.e., time period, \( t_p \), is 1 min), then there will be 10,080 \( T \) states (there are 10,080 minutes in a week).

When a user is encountered, the corresponding multi-modal biometric evidence is collected from the identifiers. An example for the biometric evidence from the identifiers and the transformed (weighted and normalised) evidence is shown in Figure 16 B in Appendix A. FR provides similarity scores, which give the percentage of similarity of the user to the known faces in the database. Age, height, and time are assumed to be discrete random variables with a discretised and normalised normal distribution of probabilities, \( N(\mu , \sigma ^2) \), defined by (1), where \( V \) is the estimated value, \( Z \) is the standard score, and \( C \) is the confidence of the biometric indicator for the estimated value. (1) \( \begin{equation} \mu = V, \quad P\left(\frac{-0.5}{\sigma } \lt Z \lt \frac{0.5}{\sigma }\right) = C \end{equation} \)

The time period and its standard deviation (\( \sigma _t \) in the normal distribution of \( T \)) can be set depending on the precision required in the application. A smaller time period and standard deviation ensure higher precision, however, this would increase the complexity of the Bayesian network, thereby increasing the time to identify the user. In addition, a higher precision carries the risks of decreasing the recognition rate if the users are not encountered near the time slot that they were previously seen.

For example, if users in the application scenario will change every 5 minutes, then \( t_p=5 \) min and \( \sigma _t=15 \) min would be reasonable. However, in an HRI scenario, \( t_p=30 \) min with \( \sigma _t=60 \) min can allow better identification, because it is less likely to encounter users around the same time every day. Hence, we use the latter in this article.

3.2 Weights of the Network

Soft biometric traits are characteristics that are not suited to identify an individual uniquely. We can assume that the population will have similar characteristics, but the distribution is unknown. However, some soft biometric features may contain more information about an individual than others, e.g., age is often more informative than gender. This can be modelled by using different weights for the parameters in a Bayesian network [45].

Weights (\( w_i \)) are used as the exponential to the likelihoods of the child nodes (\( X_i \)), similar to the work in Reference [88]. In contrast to our previous work [43], we optimise the weights of soft biometric features (gender, age, height, and time of interaction) through Bayesian optimisation, as described in Appendix C.6, while the weight of the face node (\( w_F \)) is set to be 1, as it is the only primary biometric in our system. The posterior probability \( P(I^j| X_1, \ldots , X_n) \) is approximated as in Equation (2). \( I^j \) stands for the \( j \)th user (\( I=j \)), where \( I \) is the identity node. (2) \( \begin{equation} P(I^j| X_1, \ldots , X_n) \propto \frac{P(I^j) \prod _{i} P(X_{i}|I^{j})^{w_i}}{P(X_1, \ldots , X_n)} \end{equation} \)

As in Reference [45], we assume that the identifiers perform equally well on all users. Therefore, the accuracy of an identifier is independent of the user and equal priors are assumed for each of the identifiers.

The posterior probability simplifies to Equation (3): (3) \( \begin{equation} P(I^j| X_1, \ldots , X_n) \propto P(I^j) \prod _{i} P(X_{i}|I^{j})^{w_i} \end{equation} \)

Because the distribution of users over time is not known, one approach for determining \( P(I^j) \) is to use adaptive priors using frequencies of user appearance, however, this can create a bias in the system towards the most frequently observed user, as it affects the posterior probability directly, thus, may result in a decrease in the identification rate.

Therefore, we assume that the probability of encountering user \( j \) is equally likely as encountering user \( m \), hence, we assume equal priors for \( P(I) \), as shown in Equation (4), where \( n_{e} \) is the number of enrolled users, which is updated whenever a new user is enrolled, as presented in Figure 16 in Appendix A. (4) \( \begin{equation} P(I^j) = P(I) = \frac{1}{n_{e}} \end{equation} \)

3.3 Quality of the Estimation

Algorithms for open-set problems generally use a threshold (e.g., over the highest probability/score) to determine if the user is already enrolled or “unknown.” However, the resulting posterior probabilities in a Bayesian network can be low due to the multiplication of the conditionally independent modalities and vary depending on the number of states. Hence, we use the two-step ad hoc mechanism introduced in Reference [43] to transform the Bayesian network to allow open-set recognition: (1) An “Unknown” (\( U \)) state is used in both \( F \) and \( I \) nodes. The similarity score in FR of \( U \) is set to the FR threshold (\( \theta _{FR} \)), such that when normalised, scores below/above the threshold will have lower/higher probabilities than \( U \). This allows maintaining the threshold for the FR system in use. (2) We use the confidence measure called the quality of the estimation (\( Q \)). Given the evidence \( y_t \) at time \( t \), it compares the highest posterior probability (\( P_w \)) to the second highest (\( P_s \)), as shown in Equation (5). The difference between the probabilities decreases, as the number of enrolled users (\( n_e \)) increases, since \( \sum _{j} P(I^j|y_t)=1.0 \). A similar method was used in Reference [31] for estimating the quality of localisation based on different images. (5) \( \begin{equation} Q = [P_w(I^j|y_t) - P_s(I^j|y_t)] * n_e \end{equation} \)

Using the quality of the estimation enables decreasing misidentifications. For example, the highest posterior score can be very high, but if the second highest posterior is very close to it, then it means that there are two possible strong candidates for the current user. If the system were to identify the user in this case, then the resulting misidentification could cause adverse effects on the current user especially in the case of different genders or age differences between the two users, as well as security issues. Thus, it is more preferable to identify the user as unknown if the quality is zero or below a predetermined threshold (\( \theta _Q \)) or if \( U \) has the highest posterior probability. Otherwise, the identity is estimated with a maximum a posteriori (MAP) estimation, given in Equation (6). (6) \( \begin{equation} j^* = {\left\lbrace \begin{array}{ll} U, & \text{if}\ Q = 0 \text{ or } Q \lt \theta _Q \text{ or }\\ & P(I^U|y_t) \gt P(I^j|y_t) \text{ for all } j\\ \rm{arg\,max}_j P(I^j| y_t), & \text{otherwise.} \end{array}\right.} \end{equation} \)

3.4 Incremental Learning

For personalisation in long-term HRI applications, new users may often need to be enrolled in a system to allow recognition in subsequent encounters, such as for admitting a new patient to personalised robot therapy. However, in such applications, the intermediary (e.g., clinical staff) and end-users (e.g., patients) are often non-experts, hence, systems that require the least amount of technical knowledge, effort, and time are desirable, especially those that allow users to enrol themselves. Thus, we developed an incremental learning system for the weighted multi-modal Bayesian network, which expands the network upon new user enrolment. When the MMIBN detects that the user is new, the robot requests to meet the user, and (verbally) asks for their name, gender, birth year, and height, which the user can enter through a tablet interface, after which a photo of the user is taken by the robot (step 9 in Figure 15).

This information, along with the time of interaction, is gathered to have the ground truth values for recognition and for setting the initial likelihoods of the MMIBN.

Initially, the system starts from a “tabula rasa” state, where there are no known users. Bayesian network is formed when the first user is enrolled: one state for the new user and one for the “Unknown” (\( U \)) state. Figure 16 A (in Appendix A) illustrates an example for the initial MMIBN after the enrolment of the first user, e.g., a 25-years-old female who is 168 cm tall and encountered at 11:00 am on a Monday. The initial likelihood for \( F \) is set to be much higher for the true values, as shown in Equation (7), where \( w_F \) is the weight of the face variable, and \( n_e \) is the number of enrolled users. The value was found based on preliminary experiments. (7) \( \begin{equation} P(F^k|I^j)= {\left\lbrace \begin{array}{ll} 0.9^{w_F}, & \text{if}\ k=j\\ \left[0.1/(n_e-1)\right]^{w_F}, & \text{otherwise.} \end{array}\right.} \end{equation} \)

The remaining likelihoods are set using the prior knowledge that the user entered in a similar structure to the evidence for age, height, and time variables with a discretised and normalised normal distribution, \( N(\mu , \sigma ^2) \), where \( \mu \) is the true value (e.g., age of the person), and \( \sigma \) is the standard deviation of the identifier. Gender is set at \( [0.99^{w_G}, 0.01^{w_G}] \) ratio, which is experimentally found. For the unknown state, \( P(X_i^k|I^U) \) is set to be uniformly distributed, as an unknown user can be of any age, height, and be recognised at any time of the day, except for the face node, which follows Equation (7).

When a new user is enrolled, the Bayesian network is expanded by adding a new state to \( I \) and \( F \) nodes. \( P(F^k|I^j) \) for each previous state in \( I \) (including \( U \)) is updated by appending the value corresponding to \( k \ne j \) condition in Equation (7), and then probabilities are re-normalised. The likelihoods of \( G \), \( A \), \( H, \) and \( T \) nodes for the previously enrolled users remain the same. An example of the MMIBN likelihoods during incremental learning of a new user, e.g., a 37-years-old male, 173 cm tall, and encountered on a Wednesday at 8:00 pm, is illustrated in Figure 16 E in Appendix A.

The scalability feature removes the need to retrain the network when a new user is introduced, hence, the time complexity is decreased, which can be crucial if the new user is introduced at a later step (e.g., after 1,000 users). More precisely, if each image corresponding to \( \hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt} \) average number of observations per user was to be recognised again after a new user is added to the face database, then it would take a significant amount of time to expand the network compared to scaling, since \( n_e*\hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt}*\mathcal {O}(FR) \gg n_e*\mathcal {O}(1) \) updates, where \( \mathcal {O}(FR) \) is the time complexity of the FR algorithm, and \( n_e \) is the number of enrolled users.

To reduce the risk of confusing new users with known users, it is preferable to have sufficient data within the MMIBN prior to making reliable estimations, hence, in the first few recognitions (here, we chose \( N\lt N_{min}=5 \) recognitions, i.e., the first 4 recognitions),6 the identity is declared as unknown, regardless of the estimated identity, as illustrated in Figure 16 C (Appendix A).

3.5 Online Learning of Likelihoods

Bayesian network parameters are generally determined by expert opinion or by learning from data [51]. The former can cause incorrect estimations if the set probabilities are not accurate enough. The latter, for which Maximum Likelihood (ML) estimation is commonly used, is not possible when the Bayesian network is constructed with incomplete data. One solution is to use offline batch learning, however, it requires storing data that can cause memory problems in long-term interactions. Another approach is to update the parameters as the data arrive, which is termed online learning. Variants of Expectation Maximization (EM) algorithm with a learning rate (EM(\( \eta \))) [6, 20, 57, 59] have been proposed for online learning in Bayesian networks.

We use a Bayesian network where the likelihoods are updated through EM(\( \eta \)) with an adaptive \( \eta \) (learning rate) based on ML estimation, similar to Voting EM [20]. Adopting the notation in Reference [6], the formulation is given in Equation (8). \( \theta _{ijk}^{t} \) represents the likelihood of the modality \( X_i \) at time \( t \), \( P(X_i=x_i^k|I^j) \). \( P_{\theta ^t}(x_{i}^{k}|y_{t}, I^{j}) \) represents the posterior probability of the modality \( X_i \) at time \( t \) given the current evidence \( y_t \) and the actual identity of the user \( I^j \). The difference between Voting EM and our approach is that we work with continuous probabilities due to uncertainties in the identifiers. We will refer to the proposed multi-modal incremental Bayesian network with online learning as MMIBN:OL. (8) \( \begin{equation} \theta _{ijk}^{t+1}= {\left\lbrace \begin{array}{ll} \eta _j P_{\theta ^t}(x_{i}^{k}|y_{t}, I^{j}) + (1-\eta _j) \theta _{ijk}^{t}, & \text{if}\ P(I^{j}) = 1\\ \theta _{ijk}^{t}, & \text{otherwise.} \end{array}\right.} \end{equation} \)

Combining ML estimate to achieve an adaptive learning rate (given in Equation (9)) allows the learning rate to depend on the observation of the user \( j \) (\( n_{oj} \)), which is more reliable than using a fixed rate for all users. Also, each observation of the user creates a progressively smaller update on the likelihoods, such that the effect of a new observation decreases as the number of recognitions of the user increases. (9) \( \begin{equation} \eta _j = \frac{1}{n_{oj}+1} \end{equation} \)

Supervised learning is necessary to achieve accurate online learning. The identity of the user should be known for updating the corresponding likelihoods, which can be achieved in HRI by asking for a confirmation of the estimated identity.

If the user \( j \) is previously enrolled in the system, then the likelihoods are only updated for user \( j \), as shown in Figure 16 D (in Appendix A) based on the evidence in Figure 16 B. However, if the user \( j \) is a new user, then online learning is applied on the face likelihood for the unknown state (\( P(F^k|I^U) \)), followed by incremental learning by expanding the MMIBN (as described in Section 3.4), and finally by applying online learning for the new user, as illustrated in steps 8–18 in Figure 15 and in Figure 16 F. The likelihoods of gender, age, height, and time remain the same for \( U \) to ensure uniform distribution.

3.6 Long-term Recognition Performance Loss

The standard metrics for open-set identification are Detection and Identification Rate (DIR) and False Alarm Rate (FAR) [69]. DIR is the fraction of correctly classified probes (samples) within the probes of the enrolled users (\( \mathscr{P}_\mathscr{E} \)), given in Equation (10). FAR is the fraction of incorrectly classified probes within the probes of unknown users (\( \mathscr{P}_\mathscr{U} \)), given in Equation (11). (10) \( \begin{equation} DIR = \frac{|\lbrace \rm{arg\,max}_j P(I^j| y_t) = j | j, j \in \mathscr{P}_\mathscr{E}\rbrace |}{|\mathscr{P}_\mathscr{E}|} \end{equation} \) (11) \( \begin{equation} FAR = \frac{|\lbrace \rm{arg\,max}_j P(I^j| y_t) = j | k, j \in \mathscr{P}_\mathscr{E}, k \in \mathscr{P}_\mathscr{U}\rbrace |}{|\mathscr{P}_\mathscr{U}|} \end{equation} \)

In other words, DIR represents the “true positive” (TP) of enrolled users, in which the current probe (referring to the multi-modal biometric sample) belongs to a previously enrolled user and identified correctly. FAR serves as a “false positive” (FP) for unknown users, that is, the probe belongs to an unknown user, but he/she is identified as an enrolled user. However, TP and FP are notions of verification problems, in which the probe is compared against a claimed identity, thus, are generally not applicable to open-set identification. Instead, the tradeoff between DIR and FAR that depends on the threshold of the identifier is generally represented by a Receiver Operating Characteristic (ROC) curve. The standard practice in biometric identification is to determine the desired FAR, which would then set the threshold and DIR.

Depending on the biometric application, the cost of incorrectly identifying a user as known may be very different from the cost of incorrect identification of the enrolled user [47]. For short-term interactions, in which a user will be encountered 1–2 times, FAR is as important or more important than DIR. However, for long-term interactions, users will be encountered a greater number of times. Thus, correctly identifying a user (in a closed-set) becomes more important than correctly identifying an unknown user (open-set). Hence, we introduce the long-term recognition performance loss (\( L \)) that creates a balance between DIR and FAR based on the average number of observations per user (\( \hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt} \)), as presented in Equation (12), where \( \alpha \) is the ratio of importance of \( DIR \) compared to \( FAR \).

Weights of MMIBN are optimised through this loss function, for gender, age, height, and time in \( [0, 1] \) range, along with quality (\( Q \)) that can change within \( [0, 0.5] \) range. Ideally \( L=0 \), where all unknown users are identified as such (FAR\( =0.0 \)) and the known users are correctly identified (DIR\( =1.0 \)). (12) \( \begin{equation} \begin{aligned}L &= \alpha *(1 - DIR) + (1 - \alpha)*FAR \\ \alpha &= 1 - \frac{1}{\hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt}} \end{aligned} \end{equation} \)

3.7 Normalisation Methods

The scores from each modality must be normalised into a common range (e.g., \( [0,1] \)) to ensure a meaningful combination. It is important to choose a method that is insensitive to outliers and provides a good estimate of the distribution [44], such as, minmax, tanh [37], softmax [11], and normsum (dividing each value by the sum of values). We introduce hybrid normalisation, which combines the methods that achieve the lowest loss for each modality. In other words, hybrid normalisation uses the best-performing normalisation method for each modality. Extensive tests were made on the dataset obtained from our previous work in Reference [43] to get the optimal methods for each modality (\( F \), \( G \), \( A \), \( H, \) and \( T \)). The long-term recognition performance loss was compared for each combination of the individual modality with face recognition (\( F \), \( F \)-\( G \), \( F \)-\( A \), \( F \)-\( H \), \( F \)-\( T \)) by optimising the weights for each of the combinations. The resulting hybrid normalisation uses normsum for face, gender, and height; tanh for age; softmax for time of interaction.

Skip 4RECOGNITION MODULE Section

4 RECOGNITION MODULE

While MMIBN can be applied on other platforms, its main purpose is for enabling incremental user recognition in long-term human-robot interaction in the real world. The proposed approach does not require heavy computing, therefore, it is suitable for use on commercially available robots. We employ this system on Pepper and NAO7 robots, which are amongst the most commonly used robots in HRI research [53], for our experiments (as described in Section 8). These robots are operated by NAOqi8 software, which includes different modules that allowed us to extract face similarity scores, gender, height, and age estimations from a single image through the Recognition Module in Figure 13 (Appendix A). The internal states of the proprietary algorithms (developed by OKAO) are inaccessible, hence, we assume that the gender and age estimations are not used to obtain the face similarity scores, and they are conditionally independent of the FR results, even though they are obtained from the 2D image. The height estimation in NAOqi is measured through the 3D sensor (in the eyes) of the Pepper robot and based on the face position in the 2D image and the geometric transformations (based on the camera relative to the robot) for the NAO robot. Due to relying on only one primary biometric, in the absence of facial information, the user is not recognised, since soft biometric information would not be sufficient to estimate the identity.

MMIBN can be used with any identifier software. The reason NAOqi identifiers are chosen is their capability for incremental recognition and their real-time performance, in other words, these algorithms work on a single CPU on a robot without requiring preliminary training. In contrast, the state-of-the-art deep learning methods for face recognition (such as Dlib [50]) are not optimised for low computational power systems, hence, they may require a vast amount of time for encoding images, recognition and retraining,9 which makes them unsuitable for real-time open world user recognition on a robot. Similarly, OpenFace10 [2], which is an implementation of FaceNet [79] and a popular closed-set face identification method, was found to be unsuitable for real-world HRI, because the classifier needs to be retrained after a new user enrolment with all the available data (instead of incremental learning) with batch learning of images for the new user, and the training time (albeit small) increases with the increasing number of users [2]. In addition, preliminary evaluations of OpenFace11 provided unpromising results in new user identification. For instance, the first author was recognised consistently as Anne Hathaway with a high confidence (85 to 99.2%), despite the fact that the classifier was trained on only 10 users with 600 images per user (i.e., the classifier must be very accurate in identifying known users), and the author does not resemble her that highly.

Nevertheless, it is possible to use OpenFace or other identifiers, instead of the NAOqi user recognition algorithms, for obtaining the multi-modal biometric information for MMIBN.

Skip 5MULTI-MODAL LONG-TERM USER RECOGNITION DATASET Section

5 MULTI-MODAL LONG-TERM USER RECOGNITION DATASET

Our prior work provided evidence that the proposed model is suitable for long-term HRI in the real world. However, the optimised parameters of the model could not be generalised to a larger population due to the limited number of users and their narrow age range in that study. However, collecting a diverse training set within a long-term real-world HRI scenario is very challenging. To the best of our knowledge, the only publicly available dataset that contains the soft biometrics used in our system (except for the time of interaction) with a dataset of faces is BioSoft [74]. However, due to the low number of subjects (75), and the lack of numeric height values, we decided to create our own Multi-modal Long-term User Recognition Dataset.

Datasets that contain images in the form of “mugshots,” such as NIST Mugshot Identification Database,12 do not represent real-world HRI interactions in which the obtained images from the robot’s camera may vary greatly depending on the users’ actions and the environmental conditions. Therefore, it is important to use an image dataset with real-world variations, along with ground truth values of identity, gender, and age of users to assess the performance of our model and the corresponding identifiers in similar conditions. The largest publicly available dataset of face images with gender and age labels is the IMDB-WIKI dataset [71, 72], which contains more than 500K images of 20K celebrities with a wide age range. As can be observed in Figure 3, the images in this dataset may contain bad lighting conditions, occlusions, oblique viewing angles, a variety of facial expressions, partial faces of other people, face paint and disguise, and black-and-white images, because the images come from movies, TV series, and events.

Fig. 3.

Fig. 3. Samples of images from the IMDB-WIKI dataset [71, 72] that are used in creating the Multi-modal Long-term User Recognition Dataset.

In addition to images, the estimated height of the user and the time of interaction with the robot would be necessary for user recognition in various HRI scenarios, where the users will be encountered sequentially over time. Thus, we created the Multi-modal Long-term User Recognition Dataset by (1) sampling a subset of the IMDB-WIKI image dataset and (2) artificially generating height estimations and various time of interactions to simulate repeated encounters of the users with the robot. The resulting dataset contains 200 users (101 females, 98 males, and 1 transgender person; the age range is 10 to 63) with 10 to 41 images per user that adds up to 5,735 images, height estimations, and various (patterned and random) time of interactions, along with a database of users’ names, genders, ages, and heights. Moreover, NAOqi identifier estimations (face similarity scores, gender, and age estimations) are obtained for each image and provided alongside the artificial height estimations and the time of interaction to simulate the information that would be acquired from a robot (e.g., NAO or Pepper) in an HRI scenario. The Multi-modal Long-term User Recognition Dataset is available online.13

5.1 Image Sampling

In the scope of this work, only one user is assumed to be present in each image, hence, the cropped faces of IMDB dataset is used. To simulate an open world HRI scenario, where the users will be met in consecutive days or weeks, we chose images of users that are from the same year. Furthermore, we assume that the average number of times a user will be observed is \( \hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt}\ge 10 \), which is a reasonable assumption for long-term HRI. Hence, we choose celebrities who have more than 10 images each corresponding to the same age. Moreover, to assess the incremental learning capabilities of our model with a user database that is more realistic for HRI (i.e., sufficiently large with 100 to 200 users instead of thousands of users), we (randomly) sampled 200 users out of 20K celebrities.

To create a diverse set of ages in the dataset, the images that correspond to an age that is within the five most common ages (25, 26, 28, 30, 31) in the set were randomly rejected (with 50% probability) during the selection. For instance, Anne Hathaway has sufficient images corresponding to 25 and 27 years old in the IMDB-WIKI dataset. However, 25 is among the five most common ages, thus, with a 50% chance, this set of images was excluded from the selection, hence, the images of Anne Hathaway corresponding to 27 years old were chosen instead. This also resulted in some celebrities who only have images corresponding to a certain age in the dataset to be excluded from the selection.

The resulting age range is 10–63, with the mean age of 33.04 (SD\( =9.28 \)).

Subsequently, the dataset is cleaned in three steps: by removing (1) images with a resolution lower than 150 \( \times \) 150, (2) images without a face detected by NAOqi, (3) images that erroneously correspond to another person. Furthermore, in order of user appearance (as detailed further in Section 6.2), NAOqi identifiers are applied on the selected images to obtain face similarity scores, gender, and age estimations. If the user has not been previously encountered, then the same image is used to identify the user before and after enrolment to the face database in NAOqi.

5.2 Height and Time of Interaction

Height was found to be the most important soft biometric in determining the identity in Reference [43]. To validate whether this finding persists for a large number of users with diverse characteristics and optimise its weight for applying it to real-world HRI experiments, we artificially created height data for each user.

To keep the data realistic and model the differences between the estimated heights, Gaussian noise with \( \sigma =6.3 \) cm (as found in Reference [43] for NAOqi height estimation) is added to the actual heights of the users obtained from the web.

Given our assumption that the users will be encountered at least 10 times in long-term HRI, we created two datasets: (1) D-Ten, where each user is observed precisely 10 times, e.g., 10 return visits to a robot therapist, and (2) D-All, in which each user is encountered a different amount of times (10 to 41 times). Two types of distribution are considered for the time of interaction: (1) patterned interaction times in a week modelled through a Gaussian mixture model, where the user will be encountered certain times on specific days, which applies to HRI in rehabilitation and education areas, and (2) random interaction times represented by uniform distribution, such as in domestic applications with companion robots, where the user can be seen at any time of the day in the week. As a result, we created four (sub)datasets as part of the Multi-modal Long-term User Recognition Dataset: D-TenUniform, D-TenGaussian, D-AllUniform, D-AllGaussian.

Skip 6EVALUATION Section

6 EVALUATION

In this section, we evaluate our proposed models based on the hypotheses presented in Section 6.1. The procedure of creating the cross-validation sets is described in Section 6.2. Initially, the parameters of the multi-modal incremental Bayesian network (Section 6.3) are optimised for open world recognition in long-term interactions in Section 6.4. Using those parameters, the model is compared to face recognition and soft biometrics on the Multi-modal Long-term User Recognition Dataset for the training set, closed-sets and open-set tests in Section 6.5.

6.1 Hypotheses

  • H1 Our proposed multi-modal incremental Bayesian network will improve user recognition compared to face recognition alone. As measured by a decrease in the long-term recognition performance loss (\( L \)) and an increase in the identification rate of known users (DIR).

  • H2 Online learning will improve user recognition over a non-adaptive model. As measured by a decrease in \( L \) and an increase in DIR.

  • H3 Hybrid normalisation will outperform the individual normalisation methods.

  • H4 When assumptions are made about the temporal interaction pattern of the user, recognition will improve. When the time of interaction is uniformly distributed, the loss \( L \) will be higher.

These hypotheses will be validated with various analyses, as provided in Table 1.

Table 1.
AnalysisSectionH1H2H3H4
Normalisation methodsAppendix C.5
Tukey’s HSD on lossSection 6.5.1
Tukey’s HSD on DIRSection 6.5.2
User identification in HRISection 8.1
Barista robotSection 8.2
Socially assistive robotSection 8.3
  • A check mark represents a support for the hypothesis, a cross mark represents rejecting the hypothesis, and the crossed check mark represents partial support for the hypothesis.

Table 1. The Analyses for Validating the Hypotheses and the Corresponding Results

  • A check mark represents a support for the hypothesis, a cross mark represents rejecting the hypothesis, and the crossed check mark represents partial support for the hypothesis.

6.2 Procedure

Repeated k-fold cross-validation is used to evaluate the model stability and performance. The procedure is described in Algorithm 1 in Appendix B. Two methods for creating validation folds are used, namely, OrderedKFold and ShuffledKFold. OrderedKFold is the case where users are introduced one-by-one to the system without any repetitions of previous users during the enrolment. The order of repeated interactions is random after the enrolment. In ShuffledKFold, there can be repetitions of the previous user(s) before another user is introduced, because the order of overall samples is random. OrderedKFold is similar to batch learning in an incremental learning sense, whereas the iteration (repeat) created by ShuffledKFold is more similar to a real-world scenario. Our aim is to evaluate if there are any performance differences between the two cases and to prove that the model is stable across several repeats. A stratified random bin order is used for having a different initial bin and final bin in each fold to ensure a different enrolment order of users and a different test set, respectively. We chose K\( =5 \) folds and R\( =11 \) repeats.

Each dataset (D-Ten and D-All) is divided into two with 100 users each. The first set is then divided through cross-validation procedure with 80%–20% ratio of data to the training set (first four bins, corresponding to 800 samples in D-Ten and 2,308 in D-All) and closed-set (training) (final bin, corresponding to 200 samples in D-Ten, 578 in D-All). The open-set is created from the remaining 100 users (800 samples in D-Ten, 2,280 in D-All). The closed-set (open) is similar to the closed-set (training), which corresponds to the final bin in each fold (200 in D-Ten, 569 in D-All). The open-set evaluation is made by introducing the open-set samples after the training set, that is, 100 users are enrolled in the system and recognised multiple times before the introduction of 100 new users. However, the results for the open-set do not include the results for training.

The only difference between Gaussian and uniform datasets is the time of the interaction for each sample; that is, the order of the samples is the same.

For online learning, the likelihoods are learned during the training phase (training and open-set cases), and the learned likelihoods are used without online learning for the closed-set cases.

6.3 Description of Variables

Given our datasets and the parameters of our model, we have four independent variables and three dependent variables for analysing the results on the evaluation sets: training, open-set, closed-set (training), closed-set (open). The dependent variables are DIR in Equation (10), FAR in Equation (11) and long-term recognition performance loss (shortly, loss) in Equation (12). The independent variables are as follows:

(1)

Dataset size: 10 samples per user (D-Ten), random amount of samples (D-All)

(2)

Timing of interaction: patterned interaction times (Gaussian), random interaction times (uniform)

(3)

Model: non-adaptive MMIBN, MMIBN with online learning (MMIBN:OL)

(4)

Normalisation method: softmax, minmax, tanh, normsum, and hybrid

6.4 Optimisation of Parameters

The parameters of the MMIBN need to be optimised to achieve the best recognition results. Correspondingly, we conducted several evaluations on the Multi-modal Long-term User Recognition Dataset as described in detail in Appendix C. Here, we summarise our findings for reasons of perspicuity.

Initially, the loss parameter \( \alpha \) is set as 0.9, based on our average number of observations assumption (\( \hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt}=10 \)) for long-term interaction (Appendix C.1). Subsequently, the optimum face recognition threshold with the lowest loss for (NAOqi) FR is found to be 0.4 (Appendix C.2).

MMIBN relies on the assumption that the multi-modal biometric information (face, gender, age, height, and time of interaction) are conditionally independent given the identity of the user, since the individual identifiers do not affect each other’s results. Accordingly, we assumed that the NAOqi identifiers (face, gender, and age) are conditionally independent of each other, despite relying on the same visual input (2D image). Structural learning of the Bayesian network on the Multi-modal Long-term User Recognition Dataset (in Appendix C.3) confirmed this assumption, showing that the naive Bayes classifier model is sufficient and suitable for multi-modal user identification, even when the modalities use the same input. Moreover, the average learned likelihoods in online learning are very close to the initially assumed network parameters in Section 3.4.

Bayesian optimisation14 is applied with these parameters to minimise the loss for each combination of the independent variables (40 conditions) by optimising the weights for soft biometrics and the threshold for the quality of the estimation (see Appendix C.6). Figure 4 shows how the loss decreases during the optimisation, which results in an increase in DIR at the cost of an increase in FAR. The resulting loss of MMIBN is much lower than that of FR, and correspondingly DIR and FAR are much higher. Note that \( \alpha \) can be adjusted to give more importance to FAR or a FAR can be set prior to optimisation, which may lead to a different set of optimised parameters.

Fig. 4.

Fig. 4. ROC curve for MMIBN with hybrid normalisation in the all samples dataset with Gaussian times (D-AllGaussian), with long-term recognition performance loss (Equation (12), represented with blue dots) for varying known user identification rate (DIR, represented with orange diamond shapes) and incorrect new user detection rate (FAR, x axis), for Bayesian optimisation of the weights and the quality of the estimation for 303 iterations over 5-fold cross-validation. Face recognition (FR) values are given in dashed lines (orange line representing DIR of FR, blue line for loss of FR, and black line for FAR of FR) for comparison. While optimising parameters to reduce the loss, DIR increases at the cost of increasing FAR.

While the average standard deviation of NAOqi age estimation is found to be higher (11.0) than in Reference [43] (9.3), the age is found to be the most important parameter and height the least (see Appendix C.6), in contrast with the findings in Reference [43]. Due to the higher number of users (200) and the diverse age range (10–63) in the Multi-modal Long-term User Recognition Dataset, these results are more generalisable than our prior work. Moreover, when the ground truths are not taken into account, the standard deviation of age within the estimations is found to be 8.2, which is less than the average. This is due to the appearance of users (e.g., a 30-year-old person may look like 25), which suggests that online learning of likelihoods (MMIBN:OL) may provide better recognition performance over time, as the identifiers will get better at identifying users based on their own estimations instead of ground truth values. In addition, NAOqi gender recognition is found to be equally accurate for males and females with 0.9 as the recognition rate (i.e., users’ genders are correctly recognised 90% of the time). Furthermore, using the confidence of the estimations instead of exclusively the estimated biometric data (e.g., estimated gender or age, as described in Section 3.1) allows overcoming deviations in the estimations.

With these optimised parameters, 11 repeats of 5-fold cross-validation were applied for each of the conditions (Appendix C.4), which showed that MMIBN models are stable across repeats (i.e., no significant difference in loss between repeats), and the models perform equally well for learning new users incrementally sequentially (OrderedKFold, similar to batch learning) and at random intervals (ShuffledKFold, similar to a real-world scenario). However, the size of the dataset, timing of interaction, and normalisation method are found to have significant effects on the performance of the model, however, the non-adaptive model and the model with online learning performed equally well.

Hybrid normalisation is found to outperform the other normalisation methods in all conditions (Appendix C.5), supporting our hypothesis H3. The models achieved lower loss in D-All than in D-Ten, which showed that the proposed model gets better with the increasing number of recognitions. However, hybrid normalisation with online learning (MMIBN:OL) is found to perform worse than the non-adaptive model (MMIBN), in contrast with our hypothesis H2. Moreover, most methods are found to perform significantly worse when there is no interaction pattern (uniform timing of interaction), as compared to patterned (Gaussian) interactions, supporting our hypothesis H4.

6.5 Comparison to Baselines

On the grounds that the optimised parameters of our proposed MMIBN are found, we can compare its results to face recognition (FR) and soft biometrics (SB). FR results are obtained from the NAOqi estimations by setting FR threshold (\( \theta _{FR} \)) to 0.4. SB results are obtained by giving zero weight to FR, that is, only gender and age estimates from NAOqi, artificial height estimates and time of interaction are used for identifying a user. The weights of these modalities in SB are the same as MMIBN, as shown in Figure 19 (Appendix C.6). Similarly, the weights of SB:OL are the same as those of MMIBN:OL.

We transformed a state-of-the-art open world recognition method, Extreme Value Machine15 [73] (EVM), to accept sequential and incremental data for online learning by adjusting its hyperparameters to use it as a baseline, as described in Appendix D. In the original work, batch learning of 50 classes was used with an average of 63,806 data points at each update, instead of a single data point that we used in this work. We compared our methods with the performance of two EVM models: (a) EVM:FR, using NAOqi face recognition similarity scores as data, (b) EVM:MM, using multi-modal information in the same format as it is used for our methods.

Section 6.5.1 compares the long-term recognition performance loss (shortly, loss) between the models. Appendix C.4 provides evidence that there is a significant correlation between loss and DIR, and loss and FAR, but no significant correlation is found between DIR and FAR. Hence, the analysis of loss is sufficient to determine how the model performs in comparison to others. Nevertheless, we will report the results of FAR and DIR of the models in Section 6.5.2 to further observe how the open-set recognition metrics are affected.

6.5.1 Long-term Recognition Performance Loss.

As previously mentioned, the proposed models perform better in terms of loss in D-All than in D-Ten, however, the results for D-Ten datasets show similar patterns to that of D-All. Based on the same number of recognitions for both D-All and D-Ten, which is equal to the number of samples in D-Ten for all evaluation sets, ANOVA shows that there is no significant difference in the sample size (\( p=.67 \)), as the models perform equally well for D-All and D-Ten for the same number of samples. In other words, it does not matter if each user is observed the same number of times or not. This also supports that a higher number of samples increases the performance of the models. Hence, the following analysis will only be focused on D-All, but any differences in performance between the two datasets will be noted.

We conducted Tukey’s Honestly Significant Differences (HSD) tests on the training, open-set, closed-set (training), closed-set (open) evaluation sets for D-All datasets with Gaussian and uniform timing of interaction. The corresponding plot is given in Appendix E.1.

The results show that the proposed approaches (MMIBN and MMIBN:OL) decrease the long-term recognition performance loss significantly (\( p\lt .001 \)) and substantially compared to FR, supporting the first part of our hypothesis H1. This finding is valid across all datasets (D-Ten and D-All for Gaussian and uniform times).

MMIBN performs equally well between Gaussian and uniform timing for D-All evaluation sets (i.e., no significant difference, but slightly worse in uniform), whereas it does not perform at the same significance in D-Ten evaluation sets (performs significantly worse). MMIBN:OL performance changes depending on the dataset size and the evaluation set (performs equally well only in closed-sets in D-Ten, and for training and closed-set open in D-All). Nevertheless, the models have slightly or significantly higher loss in uniform timing as compared to Gaussian, supporting hypothesis H4.

Online learning does not perform better than MMIBN, because it increases the loss at all conditions. In fact, except for training set in D-All and D-Ten and closed-sets in D-Ten for uniform timing where MMIBN and MMIBN:OL perform at the same significance level, online learning is significantly worse, which is in contrast with our hypothesis H2.

Furthermore, the results show that soft biometric features (SB and SB:OL) are not able to identify a user on their own. In general, they perform significantly worse than FR. However, when the interaction is time patterned (Gaussian), SB performs better and closer to FR as compared to uniform timing. Especially for closed-set training in D-All, it is remarkable that SB features identify the user with the same significance level performance as FR. SB and SB:OL perform mostly equally well in D-All datasets, but SB:OL performs significantly worse in several evaluation sets in D-Ten.

EVM:FR performs significantly better (\( p\lt .005 \)) than FR across all conditions. EVM:MM is significantly worse than EVM:FR (\( p\lt .01 \)) and it does not perform better than FR in most conditions. This shows that although EVM is a good method for clustering face recognition data, it does not perform well with multi-modal data.

MMIBN significantly outperforms (\( p\lt .001 \)) both EVM models across all conditions in both D-All and D-Ten. This proves that our proposed approach is significantly better than the state-of-the-art method for incremental open world recognition with multi-modal biometric information. However, EVM models use online learning instead of fixed learning rates, which could potentially lead to worse performance as observed for our model. Nevertheless, comparing EVM models to MMIBN:OL shows that MMIBN:OL significantly outperforms EVM models (\( p\lt .05 \) to \( p\lt .001 \)) in most cases, except for uniform timing for open-set and closed-set (open) in D-All and open-set in D-Ten, in which, it performs equally well with EVM:FR.

MMIBN performs equally well between training and open-set cases as well as between closed-sets, which shows that the model scales well for an increase in users (from 100 to 200 users), suggesting that the proposed approach and the optimised weights can generalise. Similar to the results in Reference [73], EVM performs equally well between those sets, showing that the change in model from batch updates to incremental updates have not changed its structure for scaling well. The models perform significantly better in closed-sets as compared to training or open-set due to the lack of unknown users in closed-sets (FAR\( =0.0 \)). Hence, loss only depends on DIR.

The models are trained on several examples of the users before the closed-set. The model performance improves with the increasing number of recognitions and stabilises towards the end (around 2,000), as can be observed in Figure 5. This supports our initial finding of performance difference between D-All and D-Ten, given that they perform equally well for the same number of recognitions. Initially, loss increases with increasing FAR when the users are introduced to the system (represented by dots in the plot). As the number of recognitions increases, the introduction of a new user does not notably increase the loss, as can be observed by the final three new users in the training set. Even though MMIBN models get better over time, they start performing consistently better than both FR and EVM models throughout both training and closed-set after only a small number of recognitions (15–48 in training, 1–6 in closed-set).

Fig. 5.

Fig. 5. The change of loss with the increasing number of recognitions for all samples dataset with Gaussian times (D-AllGaussian) for training and closed-set (training). The loss decreases with the increasing number of recognitions.

The sudden change at the beginning for the training set is due to the sequential calculation of loss for time plots: A previously enrolled person has not been identified correctly for the first time, which changes DIR from 1.0 to 0.5 (one out of two enrolled users was incorrectly identified). Note that the introduction of new users is at random order due to ShuffledKFold function, described in Section 6.2. The results for the open-set, as given in Appendix F, show a similar pattern of loss between open-set and closed-set (of the open-set cross-validation).

6.5.2 Open-set Identification Metrics: DIR and FAR.

The previously presented results confirm our claims that our proposed multi-modal Bayesian networks perform significantly better than FR, SB and EVM in long-term interactions. Nonetheless, analysing the open-set identification metrics allows us to understand how the models perform for enrolled and unknown users through DIR and FAR, respectively. The detailed presentation of Tukey’s HSD results is shown in Appendix E.2.

The results show that the increase in DIR is significant (\( p\lt .001 \)) and drastic, from 0.268 of FR to 0.657 with MMIBN and 0.561 with MMIBN:OL averaging over all the conditions in D-All (timing of interaction and evaluation set). That is a 38.9% increase in identifying the users correctly by using MMIBN, no matter the condition, which is more than double what FR is capable of providing. Hence, our hypothesis H1 that the loss will be reduced and DIR will be increased using our proposed models as compared to FR alone is fully and strongly supported.

It should be noted that the increase in DIR provided by our network is significantly higher (\( p\lt .001 \)) than DIR of soft biometrics (0.226 on average for Gaussian timing in D-All). This shows that soft biometric data are not sufficient to identify an individual, yet when combined with the primary biometric, they improve the identification rate significantly (38.9% in D-All, and 31.8% in D-Ten). This conclusion is supported by the datasets where the time of interaction is uniformly distributed (DIR of SB is 0.013 on average), that is, due to the high variability of time, the identification rate of SB is close to zero. Nevertheless, MMIBN performs equally well in Gaussian, and uniform timing within all evaluation sets in D-All, and MMIBN:OL performs equally well in D-Ten. As previously noted in H4, the loss is (slightly or significantly) higher and DIR is (slightly or significantly) lower for all datasets and MMIBN models between Gaussian and uniform timing.

MMIBN significantly outperforms both EVM methods in DIR in all datasets (\( p\lt .001 \)). EVM:FR has significantly higher DIR than FR and EVM:MM (\( p\lt .001 \)). EVM:FR performs equally well between uniform and Gaussian timing in all datasets, because it is trained only on FR data. DIR of EVM:MM drops below that of FR for uniform timing for both D-All and D-Ten, which shows that EVM is not a model to be used with time information, since the pattern of interaction with the user might not be known beforehand. Similarly, MMIBN:OL provides worse performance for uniform timing in D-All, but it always performs significantly better than or equally well with EVM:FR.

FR performs similarly in open and closed-sets in terms of loss, because it has significantly low FAR compared to MMIBN models. While low FAR is a desirable feature, the underlying reason for low FAR is that FR has very poor recognition performance on larger datasets and fails to recognise the users, because the highest similarity score returned by the identifier is lower than the threshold (\( \theta _{FR}=0.4 \)). However, as described in Appendix C.2, this threshold ensures the lowest loss for FR.

FAR of the proposed models is high because of the combination of all modalities, which increase the probability of mixing the unknown user with an enrolled user. Possible solutions to this problem will be proposed in Section 7. For our proposed models, FAR in the training set is generally slightly less than that of open-set, because of the higher number of users enrolled, but there are no significant differences across the datasets for MMIBN, supporting that the model scales well to a larger dataset without a significant decrease in performance.

In the training set, there is no significant difference between MMIBN and EVM models for FAR, and MMIBN:OL performs significantly better than EVM models for uniform timing. In contrast to MMIBN, EVM provides significantly lower FAR in open-sets than in training sets. The authors state in Reference [73] that this is due to its ability to tightly bound class hypotheses by their support.

6.5.3 User-specific Analysis.

Confusion matrices presented in Figure 6 show how users were identified throughout the training set in D-All for a fold of the cross-validation, with 0 as the ID of the unknown user and the remaining numbers corresponding to IDs of the enrolled users. The heat map represents the percentage of identification of the user as the estimated user. Ideally, the diagonal should be all dark red if users are correctly identified. However, FR (item A) mostly identifies the users as unknown, resulting in the corresponding vertical axis of 0 to be mostly red and in a low FAR and a low DIR. MMIBN (item B) has mostly red-coloured dots on the diagonal but has mixed users with other enrolled users, as can be seen from light-blue dots all over the matrix. MMIBN:OL shows a similar pattern with slight deviations.

Fig. 6.

Fig. 6. Confusion matrices of user identification for second fold of cross-validation on D-AllGaussian: (A) face recognition (FR), (B) proposed model (MMIBN), (C) incremental Extreme Value Machine (EVM) with FR data (EVM:FR), (D) incremental EVM with multi-modal data (EVM:MM). The heat map represents the percentage of identification as the estimated user. Ideally, if all users are correctly identified, then the diagonal should be dark red, and the remaining of the matrix should be dark blue.

Even though EVM:FR (item C) only uses FR information, its confusion matrix is different from that of FR. The misidentifications are highly concentrated on the final 10 users, suggesting that either FR or EVM might be subject to the catastrophic forgetting problem. Using multi-modal data overcomes that problem, as can be seen for EVM:MM (item D), as misclassifications are evenly distributed, similar to MMIBN. However, the diagonals in EVM models have notably fewer reds than MMIBN.

The significant differences in identification of users over the 5-folds of cross-validation, as presented in Appendix E.3, shows another striking result. FR does not perform equally well amongst the users in that there are significant differences of identification. Our proposed approach MMIBN balances the performance amongst users, thereby reducing any recognition bias in the system while improving the performance of the overall system significantly as compared to FR. Online learning (MMIBN:OL and EVM:FR) balances the performance further, in contrast to the decrease in performance compared to MMIBN. EVM:MM shows a similar pattern.

Figure 7 demonstrates examples from D-AllGaussian where face recognition fails to recognise the user due to the low similarity score (\( \lt \theta _{FR}=0.4 \)), whereas our proposed model identifies the user correctly based on soft biometric information. The quality of the estimation (\( Q \)) varies depending on the highest FR similarity score, as well as the disagreement between modalities. For example, for the third user (Sandra Oh), the highest FR similarity score (rank 1) is very low, corresponding to David Schwimmer, who is 28 years old in the dataset, has a height of 185 with the enrolment time of interaction on Tuesday at 18:16. Age did not provide information to differentiate the user from the incorrect estimation, whereas height and time of interaction increased the probability that the user is Sandra Oh, resulting in a correct estimation, but with a low quality score (\( 0.35\gt \theta _Q=0.013 \)). The second user (Gary Coleman) was identified correctly by FR with the highest similarity score close to, but slightly lower than \( \theta _{FR} \). This was enforced by the age estimation and the time of interaction, which compensated for the incorrect recognitions of gender and height, to get a high quality score (7.44).

Fig. 7.

Fig. 7. Examples of true values and estimated values of modalities from our Multi-modal Long-term User Recognition Dataset with Gaussian times (confidence values are given in brackets) using proposed non-adaptive multi-modal incremental Bayesian network with hybrid normalisation (referred to as BN in the figure). Highlights in red show the incorrect detection values. Face recognition was unable to recognise the users (0 represents unknown user), because the similarity scores were below the threshold (40%). Our proposed model is successful (highlighted in green) in correctly identifying the users with varying quality of estimations (shown in brackets underneath the ID) as a result of the information gathered from soft biometrics highlighted in blue. 8% confidence value of height corresponds to the \( \sigma =6.3 \) cm in NAOqi.

6.5.4 Real-time Capabilities.

In contrast to the state-of-the-art deep learning methods, the proposed models can run on a commercial robot with low computational power (on a single CPU of Pepper robot) and only require a small amount of time for execution. In addition to the time required from FR and other modalities (M\( =0.14 \) s, SD\( =0.001 \)), MMIBN models take 0.01 second for recognition, significantly outperforming both EVM:FR and EVM:MM, which take 0.32 and 0.34, respectively.16 For enrolling new users, MMIBN requires a significantly lower amount of time (0.39 s, \( p=.002 \)) for scaling the Bayesian network, compared to MMIBN:OL, which takes 0.54 s, for which 0.17 s is due to online learning. There is no significant difference between MMIBN:OL and EVM models for enrolling (EVM:FR takes 0.48 and EVM:MM takes 0.52 s), with 0.20 and 0.23 s for online learning, respectively. The higher amount of time required for EVM:MM compared to EVM:FR shows that online learning takes longer time when there is more information to be learned per user. Note that the time required for MMIBN has decreased from 0.3 s in Reference [43] to 0.01, as a result of optimising the MMIBN algorithm.

Moreover, in comparison to deep learning approaches, which require “big data” to be pretrained, our proposed models are able to start from a state of no enrolled users, learn users continuously and incrementally, and improve performance compared to FR after a small number of recognitions (e.g., 48 in Figure 5).

Skip 7DISCUSSION Section

7 DISCUSSION

Our findings showed that from our initial hypotheses H1 and H3 are fully supported, H4 is supported for hybrid normalisation, and H2 is rejected (Table 1). In this section, we will discuss the implications of our results, validate our assumptions, and offer other approaches for our models.

7.1 Dataset Size

In general, the loss is lower, DIR and FAR are higher in D-All than in D-Ten. The increase in DIR and the decrease in loss can be explained by the higher number of recognitions, which increases the performance over time. The increase in FAR can be due to different optimised weights for each dataset (see Figure 19 in Appendix C.6). However, both datasets show similar patterns in differences between FR, SB, and MMIBN models. Even though the number of samples per user is not the same in D-All, the fact that it performs equally well as D-Ten for the same number of recognitions shows that our equal priors assumption (Equation (4)), which states that each user is equally likely to be seen, does not have any adverse effect on our proposed models.

While the weights of the biometric information differ based on the dataset size and the model, their positive values indicate that each modality is beneficial and effective in identifying users, and conditionally independent of each other, as supported by the learned structure (Appendix C.3). We suggest using the optimisation parameters (weights and quality threshold) that are optimised for D-All datasets, since this dataset contains more samples. If the application is based on users appearing at specified times during a week (e.g., long-term therapy in a hospital), then the optimised parameters for D-AllGaussian should be used; otherwise, it is better to use that of D-AllUniform (e.g., for companion robots). These optimised parameters generally perform significantly equivalent in both timing conditions in D-All for both models, as shown in Figure 20, even though the timing of interaction does not provide enough information in the uniform timing case. Nonetheless, using different (or more accurate) identifiers for soft biometrics may result in a different set of weights and better recognition performance.

7.2 High False Alarm Rate

High FAR of the models is due to the tradeoff between recognition and spotting unknown people, which is visible in Figure 4. The value of \( \alpha \) determines the importance of this tradeoff in the loss function to ensure a higher number of correct recognitions in a long-term interaction. We found \( \alpha =0.9 \) based on our assumption, that the average number of interactions is 10. Using a varying amount of samples (D-All) did not change the overall performance in terms of long-term recognition loss for the same number of total samples when we compared D-All and D-Ten at the same amount of samples (800 for training and open-set and 200 for closed-sets). In Figure 5, 71% of the users had less than 10 recognitions and 20% had more than 10, before the 800th recognition in D-All dataset. This finding shows that our choice of \( \alpha \) did not negatively affect the results. Thus, instead of changing \( \alpha \) for decreasing FAR, we would suggest using a variable threshold of quality (\( \theta _{Q} \)) based on the number of users in the dataset to ensure that the quality is higher when the number of users is low.

The presented results are dependent on the noise level of the identifiers and the characteristics of the population (e.g., the distribution of parameters within the population). By using other algorithms for the identifiers or by setting a desired FAR depending on the application from Figure 4, a different set of weights can be achieved with lower/higher FAR and consequently lower/higher DIR.

7.3 Online Learning

We initially assumed that all identifiers work equally well on all users based on the work in Reference [45]. However, there can be changes in the person’s appearance, the similarity between users, as well as changes in time of interaction, which could negatively affect the visual identifiers and the time component of our models, respectively. We claimed that our online learning approach would adjust to these changes and perform better than the non-adaptive model (H2), but the second part of the hypothesis is not supported, because online learning (MMIBN:OL) performed significantly worse or at the same significance as the non-adaptive MMIBN.

The underlying reason might be the accumulating noise in the identifiers. We suggest three possible solutions for improving online learning: (a) identifiers with lower noise can be used, which can be difficult to achieve in real-world scenarios; (b) similar to the work in References [20, 59], the learning rate \( \eta \) can be increased when there is a large error between the estimated parameter and its mean value, and decreased when convergence is reached; (c) confidence value of the identifiers or the quality of the estimation can be used to determine if the likelihoods should be updated at each iteration to avoid updating when the noise is high. However, the average learned likelihoods in online learning showed that the initial parameter assumptions in Section 3.4 hold valid. Online learning can also be applied to the weights of the MMIBN nodes to improve recognition performance over time based on the identifier accuracy, through decreasing or increasing the weights of the identifiers that are less or more accurate based on the data. We suggest applying online learning, similar to References [40] or [56], on top of the optimised weights found in this work, which would allow adapting the MMIBN to work equally well (or better) with any (i.e., NAOqi or other) identifiers. However, a simpler approach is to apply Bayesian optimisation (of the weights and the quality of the estimation) on the Multi-modal Long-term User Recognition Dataset before deploying the MMIBN with other identifier algorithms to the real-world HRI applications.

FR does not perform equally well on users, as shown in Appendix E.3. Our proposed MMIBN models decrease the recognition bias in the system using multi-modal information. This finding is also confirmed for the uniform timing of interaction. Moreover, the first part of our hypothesis that online learning will adjust to these changes is supported, which allowed decreasing the bias of FR further. We can conclude that for long-term recognition our multi-modal incremental Bayesian networks not only perform better than FR alone in all datasets but also increases performance on each user to identify them equally well.

Skip 8USER RECOGNITION IN LONG-TERM HUMAN-ROBOT INTERACTION IN THE REAL WORLD Section

8 USER RECOGNITION IN LONG-TERM HUMAN-ROBOT INTERACTION IN THE REAL WORLD

8.1 User Identification Study

In our prior work [43], we proposed and applied a multi-modal weighted Bayesian network with online learning (MMIBN:OL) to a long-term HRI scenario (through the recognition architectures in Appendix A), where 14 participants (4 female, 10 male, of age range 24–40) interacted with the robot for four weeks in an office at the University of Plymouth (Figure 8). The video showing the interaction for a known user is available online.17 The study showed that our proposed approach enables and facilitates incremental identification in a real-world HRI scenario. Moreover, the optimised parameters on the real-world data showed an improvement (1.4% increase in DIR for closed-set and 4.4% in open-set) over face recognition (DIR\( =0.903 \)). Furthermore, MMIBN using minmax normalisation (MMIBNMinmax) and MMIBN:OL with softmax (MMIBN:OLSoftmax) were the best-performing methods on the data using zero weights on age and time of interaction. However, the resulting dataset was limited in terms of the number of participants and the characteristics of the participants, hence, the results and the optimisation parameters could not be generalised.

Fig. 8.

Fig. 8. A user is interacting with the Pepper robot (SoftBank Robotics Europe) to confirm the identity that is estimated, during the user identification study in Reference [43].

Correspondingly, we created the Multi-modal Long-term User Recognition Dataset to optimise the parameters of our models and validate them on a large number of users in varying conditions with a high variability of subject age and heights, which are highly challenging to obtain in an HRI experiment. The previous sections provided conclusive evidence that our proposed models are suitable for long-term user recognition, generalise well to new users and provide significantly more reliable identification than the state-of-the-art open world recognition model (Extreme Value Machine) and (NAOqi) face recognition alone. This section evaluates how the baselines and the optimised models in this work performed on the raw HRI data in comparison to the models in Reference [43].

McNemar test is the best statistical method for comparing two classification algorithms that are run only once [28]. Cochran’s Q test is an extension of the McNemar test for more than two groups. Thus, Cochran’s Q test is applied to compare the identification of enrolled users (i.e., DIR) and new users (FAR) of all methods separately, and pairwise McNemar using Benjamini-Hochberg adjustment for multiple comparisons is applied as the post hoc test [60]. The results show that there is a significant difference between all models (\( p\lt .001 \), \( Q=161.44 \), df\( =6 \)) for DIR and the pairwise comparisons are shown in Table 2. MMIBN models with optimised parameters on the D-AllUniform dataset are used as the users were randomly encountered, however, no significant differences are observed between the models that were trained on the D-AllGaussian dataset (not shown for brevity). The results confirm that the optimised models in this work perform equally well as those that were optimised on the real-world data when the learning method is the same (e.g., comparing online learning models). Moreover, all MMIBN models significantly outperform FR (DIR\( =0.881 \), L\( =0.127 \)) (supporting hypothesis H1) and EVM:MM (DIR\( =0.858 \)). Furthermore, the losses of MMIBN models are less than FR after only 39 recognitions. While the DIR of MMIBNMinmax is slightly higher (DIR\( =0.932 \), L\( =0.135 \)) than MMIBNUniform (DIR\( =0.929 \), L\( =0.117 \)), MMIBNUniform has the lowest loss. Similar to the previous results, online learning does not outperform the non-adaptive model, in contrast to our hypothesis H2. EVM:FR does not perform significantly different than the MMIBN models, however, it does not reach their performance over time (Figure 9). Moreover, EVM models take substantially higher time to identify users (0.12 s for EVM:FR and 0.13 s for EVM:MM) than the MMIBN models (0.01 s for MMIBN and 0.03 s for MMIBN:OL). While there does not exist significant differences between the models in terms of FAR due to the low number of enrolled users, FR performs best (FAR\( =0.2 \)), followed by MMIBNUniform and EVM:MM (FAR\( =0.53 \)).

Fig. 9.

Fig. 9. Model performance on (A) known user identification (DIR) and (B) new user (incorrect) detection (FAR) on the raw data (1,272 samples) from the user identification in HRI study [43] over four weeks for the optimised models in Reference [43] (MMIBNMinmax and MMIBN:OLSoftmax), baselines (FR, EVM:FR, EVM:MM), and the optimised MMIBN models on the D-AllUniform dataset.

Table 2.
ModelFREVM:FREVM:MMMMIBN MinmaxMMIBN:OL SoftmaxMMIBN Uniform
EVM:FR\( p\lt .001 \) (\( Z=-4.91 \))-----
EVM:MM\( p=0.06 \) (\( Z=2.06 \))\( p\lt .001 \) (\( Z=6.67 \))----
MMIBN Minmax\( p\lt .001 \) (\( Z=-7.25 \))\( p=.57 \) (\( Z=-0.742 \))\( p\lt .001 \) (\( Z=-6.99 \))---
MMIBN:OL Softmax\( p\lt .001 \) (\( Z=-6.47 \))\( p=.75 \) (\( Z=0.404 \))\( p\lt .001 \) (\( Z=-6.15 \))\( p=.04 \) (\( Z=2.29 \))--
MMIBN Uniform\( p\lt .001 \) (\( Z=-6.63 \))\( p=.76 \) (\( Z=-0.308 \))\( p\lt .001 \) (\( Z=-6.62 \))\( p=.14 \) (\( Z=1.63 \))\( p=.23 \) (\( Z=-1.35 \))-
MMIBN:OL Uniform\( p\lt .001 \) (\( Z=-5.74 \))\( p=.64 \) (\( Z=0.606 \))\( p\lt .001 \) (\( Z=-5.87 \))\( p=.01 \) (\( Z=2.71 \))\( p=.75 \) (\( Z=0.365 \))\( p=.07 \) (\( Z=1.96 \))
  • Significant differences (\( p\lt .05 \)) are highlighted in bold.

Table 2. Pairwise McNemar Test Results on the Identification of Known Users (DIR) for Raw User Identification Data from Reference [43]

  • Significant differences (\( p\lt .05 \)) are highlighted in bold.

8.2 Personalised Barista Robot

In a typical coffeehouse, baristas serve hundreds of customers per day and would not be able to recognise return customers or recall their preferences. A personalised robot could recognise a high number of customers, refer to them by name and recall and recommend their favourite orders, which could improve the customer experience and reduce the order time. In such an application, the customers will arrive sequentially at random times, and they need to be autonomously and incrementally added to the system with minimum time and effort from the customer. MMIBN corresponds to these requirements for incremental long-term user recognition in real-time. Consequently, the non-adaptive MMIBN with the optimised parameters on the D-AllUniform dataset was applied for identifying customers with a personalised barista robot (using the Adapted Pepper18 robot) that recalls customer preferences [42].

A five-day HRI study with a generic (non-personalised) and a personalised barista robot was conducted in the coffee bar of an international student campus, Cité Internationale Universitaire de Paris (France), with 18 non-native English speakers (11 male, 7 female) within the age range of 22–47 (Figure 10). Speech recognition was used to make the interaction more natural, and the confirmation of the estimated identity was implicitly taken through the dialogue (i.e., if the user does not oppose the estimated identity, the identity was assumed to be correct), in contrast to Reference [43] where the user needed to explicitly confirm the identity through the tablet interface (as shown in Figure 8). Also, ground truth values (gender, age, height, and an explicitly taken image) were not requested to reduce the effort required by the customer (i.e., step nine in Figure 15 was not used), thus, only the estimated values were used for enrolling users. However, users either did not realise that the estimated identity was incorrect or the identity was incorrectly confirmed due to speech recognition errors,19 which resulted in a high FAR (FAR\( =0.786 \) for MMIBN, FAR\( =0.286 \) for FR) and prevented some of the new users to be enrolled, showing the necessity of explicit user confirmation. Nonetheless, MMIBN performed better (DIR\( =0.75 \), L\( =0.304 \)) than NAOqi FR (DIR\( =0.5 \), L\( =0.479 \) for 12 known user recognitions), supporting our hypothesis H1. Moreover, personalisation was found to mitigate the negative user experience, which suggests that user recognition plays an important role in long-term HRI. On average, 3.1 seconds (SD\( =0.9 \)) were taken to recognise users, which includes the time for user detection and the recognition module (Figure 13) to obtain the biometric samples and the time for MMIBN to identify the user (0.01 s).

Fig. 10.

Fig. 10. Personalised barista robot [42] at the coffee bar of an international student campus, Cité Internationale Universitaire de Paris (France).

8.3 Personalised Socially Assistive Robot

Another area where personalisation can have an impact on long-term HRI is rehabilitation. Previous research shows that personalising the therapy improves user motivation and engagement, helps clinical staff in monitoring the progress of the patient, and facilitates rapport and trust over long-term interactions [19, 70, 75, 84]. Such improvements are desirable to improve adherence in cardiac rehabilitation, which is a long-term programme offered to those who suffered a cardiovascular event to accelerate recovery and reduce the risk of suffering recurrent events through structured exercise, education, and risk factor modification [35, 52]. Thus, in collaboration with medical specialists, a personalised socially assistive robot and a sensor interface [14, 15, 41, 54] were designed and deployed for long-term (18 weeks) cardiac rehabilitation programme at the Fundación Cardioinfantil-Instituto de Cardiología (Bogotá, Colombia), as shown in Figure 11, for five months before the outbreak of COVID-19 (which halted the programme at the clinic in March 2020). Because the robot is deployed in rehabilitation with non-expert users (e.g., doctors, nurses, patients), it should be autonomous and require minimal effort from users and medical staff [30]. Accordingly, an incremental user recognition system that does not need preliminary training is necessary for personalisation of the interaction, thus, MMIBN was chosen as the user recognition method. However, because the users will be generally encountered at patterned times (i.e., at their appointments twice per week), MMIBN with online learning with the optimised parameters on the D-AllGaussian dataset (MMIBN:OLGaussian) was used to evaluate its performance in a real-world interaction.

Fig. 11.

Fig. 11. Personalised socially assistive robot (using the NAO robot from SoftBank Robotics Europe) for long-term cardiac rehabilitation programme [14, 15, 41, 54] at the Fundación Cardioinfantil-Instituto de Cardiología (Bogotá, Colombia).

In contrast to the previous experiment [42], we used explicit confirmation of identity, in addition to the ground truth values for user enrolment, to avoid errors. The average recognition response time, which includes user detection, estimation of biometrics and identity, request of identity confirmation, the confirmation by the user on the tablet interface, and the updating of the model parameters (steps 1 to 10 in Figure 15), was 24.8 seconds (SD\( =15.5 \)) for known user recognition, and 83.6 s (SD\( =39.3 \)) for new user enrolment, including the user to enter the ground truth values on the tablet (steps 1 to 18). Considering that the system is used by non-experts (patients), the time required is not substantial, especially because the patients take on average 9.39 s (SD\( =17.46 \)) to give a response to the tablet. MMIBN:OL took 0.04 s (SD\( =0.01 \)) for recognition.

Figure 12 shows the performance of MMIBN:OL over time (with the increasing number of recognitions), and the performance of the other models on the real-world data is presented for comparison. Thirteen patients participated in the cardiac rehabilitation programme with the personalised robot, however, as observable from the figure, 30 enrolments were made to the system. The reason was a recurrent NAOqi face recognition failure that was never experienced in any of the prior studies, which resulted in erroneous user enrolments without registering the user’s image to the face recognition database, thus, DIR dropped considerably. The experimenters at the hospital addressed the issue by re-enrolling some of the patients as new users, and the issue was resolved completely after the study by adding a threshold (e.g., 0.4) on NAOqi face recognition confidence. Nonetheless, Cochran’s Q test shows significant differences between all models (\( p\lt .001 \), \( Q=21.49 \), df\( =4 \)) for identifying enrolled users. Table 3 shows that there are significant differences between the MMIBN models and FR (DIR\( =0.34 \), L=0.61), in addition to FR and EVM:MM. In contrast to our results in Section 6.5, MMIBN:OL performed slightly better than MMIBN in identifying known users (DIR\( =0.38 \) for MMIBN:OL, DIR\( =0.36 \) for MMIBN), notably better in identifying new users (FAR\( =0.56 \) for MMIBN:OL, FAR\( =0.67 \) for MMIBN), and achieved lower loss (L\( =0.62 \) for MMIBN:OL, L\( =0.64 \) for MMIBN), supporting our hypothesis H2, however, no significant differences are observed between the models. However, FR performed significantly better in FAR (FAR\( =0.13 \), \( p\lt .001 \)) than all baselines, because it identified most (63%) of known users as new. Because of the lower FAR and improving FR with re-enrolments, FR achieved a slightly lower loss than MMIBN:OL after 260 recognitions, thus, providing only partial support for our first hypothesis (H1).

Fig. 12.

Fig. 12. Model performance on (A) known user identification (DIR) and (B) new user (incorrect) detection (FAR) throughout the cardiac rehabilitation programme with the personalised socially assistive robot, lasting five months (535 recognitions). MMIBN with online learning using optimised weights on the D-AllGaussian dataset (MMIBN:OLGaussian) was used for user identification during the programme. The performance of the other models on the real-world data are presented here for comparison.

Table 3.
ModelFREVM:FREVM:MMMMIBNGaussian
EVM:FR\( p=.44 \) (\( Z=-0.853 \))---
EVM:MM\( p\lt .001 \) (\( Z=-4 \))\( p=.02 \) (\( Z=-2.8 \))--
MMIBNGaussian\( p=.02 \) (\( Z=-2.56 \))\( p=.68 \) (\( Z=-0.413 \))\( p=.02 \) (\( Z=2.68 \))-
MMIBN:OLGaussian\( p=.02 \) (\( Z=-2.65 \))\( p=.44 \) (\( Z=-0.971 \))\( p=.07 \) (\( Z=2.01 \))\( p=.44 \) (\( Z=-0.849 \))
  • Significant differences (\( p\lt .05 \)) are highlighted in bold.

Table 3. Pairwise McNemar Test Results on the Identification of Known Users (DIR) for the Socially Assistive Robot Study

  • Significant differences (\( p\lt .05 \)) are highlighted in bold.

While EVM:MM performs best overall in DIR and loss (DIR\( =0.42 \), FAR=0.67, L=0.57), EVM:FR performs the worst of all models (DIR\( =0.36 \), FAR\( =0.8 \), L\( =0.66 \)), which is in contrast with the findings in Sections 6.5 and 8.1. Moreover, users were not recognised for the first 29 recognitions with EVM because of its tail size parameter that was optimised on the multi-modal dataset, and lowering it gave erroneous results. In contrast, only the first four estimations of MMIBN are discarded (i.e., users were identified as new, regardless of the model estimation), as in the multi-modal dataset. Furthermore, EVM models take 0.12 s for user recognition, which is substantially higher than MMIBN models (0.01 for non-adaptive model and 0.04 with online learning). These findings further support that MMIBN models are the most reliable state-of-the-art open world user recognition method for HRI.

Overall, our findings on the Multi-modal Long-term User Recognition Dataset and the real-world HRI experiments show that both of our proposed approaches perform better in recognising users than the state-of-the-art open world recognition method (Extreme Value Machine) and the NAOqi face recognition alone, supporting that our proposed user recognition models are suitable for incremental user identification in real-world HRI, and that they improve the recognition even when the identifiers are malfunctioning.

Skip 9CONCLUSION Section

9 CONCLUSION

User identification is mostly regarded as a solved problem in the computer vision field. What remains unsolved is its application to the real world on low-computational power systems, such as commercial robots. The core problem that we face within HRI for personalising the interaction is to recognise unknown users and enrol them incrementally, which is classified as open world recognition. However, there exists a limited amount of research on this topic, and none of the available methods is evaluated on user identification. These methods use batch learning of classes instead of sequential learning, which is unlikely to be the case for HRI, because the users might not be available at the same time. In contrast, it is more likely that the same users will be encountered several times before the introduction of another.

Moreover, the computer vision field is not generally concerned with long-term interactions. Hence, correct identification of the enrolled users (DIR) and incorrect identification of the unknown users (FAR) are of equal value, whereas the former is more valuable in long-term interactions, since the same user is expected to be recognised several times, and the fraction of newly enrolled users will be much less. Furthermore, the appearance of the user may change over time, which requires updating the user database accordingly through online learning. In addition, combining soft biometrics, which are ancillary physical or behavioural characteristics (e.g., age) that can be extracted from primary biometric data (e.g., face) or available through other sources of information (e.g., time of interaction), can improve recognition accuracy.

In this work, we addressed these open challenges and presented a multi-modal incremental user recognition approach with online learning that is suitable for long-term HRI in the real world. We validated the approach within a variety of settings using an artificially generated multi-modal dataset and through three real-world HRI experiments, thereby extending the findings in our prior work [43] for a large number of users.

APPENDICES

A RECOGNITION ARCHITECTURE

The recognition architectures presented in Figures 13, 14, and 15 were used for the HRI experiments described in Section 8, namely, user identification in a research office [43], the personalised socially assistive robot for cardiac rehabilitation [41], and the personalised barista robot [42], as well as for evaluations on the Multi-modal Long-term User Recognition Dataset (Section 6). The Recognition Module (Figure 13) for NAOqi proprietary software was used to obtain the face similarity scores and gender, age, and height estimations, along with the time of interaction, however, the last two parameters were artificially generated for the multi-modal dataset, as described in Section 5. The identifiers in the Recognition Module can be replaced with any software providing the same biometric estimations. The image, estimated and true identity, and ground truth values are automatically and incrementally fed into the system for the multi-modal dataset; in contrast, the image is taken (via the camera on the robot’s or the tablet) when a user arrives, the estimated identity was announced to the user by a robot and confirmed by the user, and the ground truth values are entered by the user (through a tablet interface) in the HRI experiments. Figure 16 illustrates user estimation and how the prior and likelihoods of the MMIBN change for incremental and online learning based on known and new users and the evidence from the identifiers.

Fig. 13.

Fig. 13. Diagram of the recognition module. The yellow highlighted modules are proprietary software within NAOqi that are used to obtain the estimated modalities.

Fig. 14.

Fig. 14. Diagram of the estimation of the identity within the Multi-modal Incremental Bayesian Network (MMIBN). \( N \) is the number of the recognition, \( N_{min} \) is the minimum number of recognitions to ensure an identity is estimated correctly by MMIBN (taken as 5, as explained in Section 3.4), \( Q \) is the quality of the estimation in Equation (5) that is compared to the threshold ( \( \theta _{Q} \) ), \( P(I^j|y_t) \) is the posterior of the identity for user \( j \) and unknown user \( U \) in Equation (6).

Fig. 15.

Fig. 15. Diagram of the recognition architecture for Multi-modal Incremental Bayesian Network with Online Learning (MMIBN:OL). Steps 8 and 16 are not used in non-adaptive MMIBN. The first 7 actions in the architecture are common to both known (enrolled) and new users. Actions 8–18 (in green) are performed for new users, whereas 8–10 (in red) are performed for known users. Dashed line shows that ground truth values for gender, age, and height are requested from the user and a new image is taken as input when the user is enrolling (if this step is skipped, estimated values will be used for enrolment). The recognition module and the estimation of the identity within the MMIBN are presented in Figures 13 and 14, respectively.

Fig. 16.

Fig. 16. Illustration of MMIBN: (A) initial model after the enrolment of the first user ( \( j=1 \) ), (B) user evidence from identifiers and transformed (weighted and normalised) evidence during user recognition, (C) estimated user ( \( j^* \) ) based on posterior score ( \( P(I|F,G,A,H,T) \) ), the minimum ( \( N_{min} \) ) number of recognitions ( \( N \) ), and quality of the estimation ( \( Q \) ) and its threshold ( \( \theta _Q \) ), (D) model with online learning for a known user ( \( j=1 \) ), (E) model with incremental learning for a new user ( \( j=2 \) ), (F) model with incremental and online learning for a new user ( \( j=2 \) ). Non-adaptive model remains the same as in (A) for a known user ( \( j=1 \) ). Updated likelihoods for \( I \) , \( F \) , and \( G \) are shown in red for incremental learning, and blue for online learning. For \( A \) , \( H, \) and \( T \) nodes, \( j=0 \) (unknown) is represented by light blue, \( j=1 \) by orange, and \( j=2 \) by green.

B REPEATED K-FOLD CROSS-VALIDATION GENERATION

As described in Section 6.2, two methods are used to create the cross-validation repeats, as presented in Algorithm 1: OrderedKFold, where users are enrolled one after another and the enrolment order is different in each fold, and ShuffledKFold, where the user samples (probes) are shuffled, hence, users may be repeatedly seen before another user is enrolled.

C OPTIMISATION OF PARAMETERS

Initially, the loss parameter \( \alpha \) and face recognition threshold is set as described in Sections C.1 and C.2. Furthermore, structural learning is applied to the data to validate the assumption of conditional independence in the Bayesian network (Section C.3). Subsequently, Bayesian optimisation is used to optimise the weights of the network and the threshold for the quality of the estimation (\( \theta _Q \)). A total of 303 iterations is used for 5-fold cross-validation for each combination of the independent variables (for 40 conditions). The parameters are optimised by minimising the loss on the training set. By using the optimised parameters, 11 repeats of 5-fold cross-validation are conducted for each of the conditions to evaluate the effects of the independent variables on the open-set. For clarity of the presentation of results, we will initially analyse the results for 11-repeats of 5-fold cross-validation (in Section C.4) before presenting the optimised parameters from Bayesian optimisation. This would allow us to later analyse only the optimisation parameters (Section C.6) for the best-performing normalisation method (Section C.5).

Skip C.1Loss Parameter Section

C.1 Loss Parameter

The loss parameter \( \alpha \) (Equation (12)) should be set to find the optimum FR threshold (\( \theta _{FR} \)) and optimise the parameters in our network. As \( \alpha \) increases, the fraction of correct recognitions of enrolled users (DIR) increases, but the fraction of the incorrect recognitions of unknown users (FAR) will increase. Based on our average number of observations assumption for long-term interaction (\( \hspace{0.83328pt}\overline{\hspace{-0.83328pt}n_o\hspace{-0.83328pt}}\hspace{0.83328pt}=10 \)), \( \alpha \) becomes 0.9.

For applications with fewer observations per user, \( \alpha \) can be set accordingly.

Skip C.2Face Recognition Threshold Section

C.2 Face Recognition Threshold

In FR, if the highest similarity score is below the face recognition threshold, \( \theta _{FR} \), then the identity is classified as unknown. We examined how \( \theta _{FR} \) influences the long-term recognition performance loss for the NAOqi FR in both D-Ten and D-All datasets and noticed a decrease in performance (i.e., increase in loss) for \( \theta _{FR}\gt 0.4 \). Hence, we chose \( \theta _{FR}=0.4, \) because it is the highest threshold giving the lowest loss to decrease FAR in our model, in agreement with our previous work in Reference [43].

Skip C.3Bayesian Network Structure Section

C.3 Bayesian Network Structure

To determine whether the conditional independence of the modalities (face, gender, age, height, and time of interaction) given the identity of the user holds when the same input (i.e., image) is used to obtain multi-modal data, we applied structural learning of the Bayesian network on the Multi-modal Long-term User Recognition Dataset using the pyAgrum [36] library. We used the all samples dataset with Gaussian times (D-AllGaussian) with the identification estimations obtained from NAOqi proprietary algorithms (for face, gender, and age estimations) and the artificially generated height estimations and time of interactions, as described in Section 5. Based on the requirements of the pyAgrum library, the multi-modal data is “simplified” by taking the best match evidence for modalities (i.e., confidence scores are not used) to allow structural learning. For instance, the most similar user (or unknown) is taken as the face recognition estimate by taking into account the face recognition threshold, and the evidence for gender, age, and height are taken as the estimated values. Mandatory arcs (e.g., \( I \) -> \( F \)) between the identity node and the modalities are provided as prior structural knowledge, since the multi-modal information is used to determine the identity. Based on the Bayesian Dirichlet equivalent uniform (BDeu) score [39], all three methods available in the library (K2 algorithm [22], greedy hill-climbing search, and local search with tabu list) found no other dependencies between the modalities, confirming the conditional independence.

We initially set the likelihoods to have much higher values for the true values, such as 0.9 for the face node (Equation (7)) corresponding to the actual user and \( 0.99^{w_G} \) for the true gender. Average learned likelihoods in online learning for 200 users confirm this assumption, with the mean for face node as 0.913 (SD=0.126), and the mean for the gender likelihood as 0.978 (SD=0.058).

Skip C.4Analysis of Variance of Independent Variables Section

C.4 Analysis of Variance of Independent Variables

Levene’s test on the loss reveals (\( F(10,2189)=0.026, p=1 \)) that there is no significant difference in variances between the repeats, which indicates that our models are stable across repeats. ANOVA (Type-I) supports that there is no significant difference between repeats (\( F(10,2189)=0.044, p=1 \)), which shows that there is no significant difference between the ordered k-fold cross-validation and the shuffled k-fold, indicating that the model performs equally well for learning new users incrementally sequentially (similar to batch learning) and at random intervals (similar to a real-world scenario). Hence, we will only analyse the results of a single randomly selected repeat of 5-fold cross-validation. Since the model is stable across repeats, using a single repeat of cross-fold validation instead of independent test sets does not violate ANOVA assumption [7].

Due to the linear relation of loss with DIR and FAR in Equation (12), there will be a correlation between the parameters. Pearson’s product-moment partial correlation coefficient was computed to assess their relationships. The results show that there is a negative correlation between loss and FAR, \( r(200)=-0.18, p=.009 \), a positive correlation between loss and DIR, \( r(200)=0.99, p\lt .001 \), but no significant correlation between FAR and DIR, \( r(200)=0.08, p=.25 \).

A factorial ANOVA is conducted for analysing the primary and interaction effects of our independent variables. The results show that there are no significant primary effects for the model, \( F(1,160)=1.50, p=.22 \), and no significant interaction effects are found between the dataset size, timing of interaction, and model combination, \( F(1, 160)=0.01, p=.91 \). Every other independent variable and their interactions are found to be significant (\( p\lt .001 \)). This shows that the size of the dataset, timing of interaction, and normalisation method have significant effects on the performance of the model, but online learning by itself does not provide significant improvement.

Skip C.5Normalisation Methods Section

C.5 Normalisation Methods

A post hoc analysis using Tukey’s Honestly Significant Differences (HSD) test was conducted in which D-All and D-Ten datasets have been analysed separately for clarity, however, the results show similar patterns in both datasets. The corresponding Tukey’s HSD test plots are presented in Figures 17 and 18 (see Appendix E for the description of significance levels).

Fig. 17.

Fig. 17. Results of Tukey’s HSD test of loss in the open-set for normalisation methods with optimised weights for all samples (D-All) dataset: softmax, minmax, tanh, normsum, and hybrid. Lower loss is better.

Fig. 18.

Fig. 18. Results of Tukey’s HSD test of loss in the open-set for normalisation methods with optimised weights for 10 samples (D-Ten) dataset: softmax, minmax, tanh, normsum, and hybrid. Lower loss is better.

In both of D-All and D-Ten datasets, hybrid normalisation provides significantly lower loss (\( p\lt .05 \)) in all conditions except for online learning in Gaussian timing for D-All (\( p=.78 \) in D-AllGaussian), in which case it still provides the lowest mean for loss. Hence, our hypothesis H3 is strongly supported, and the hybrid normalisation method is chosen for the remaining analyses.

While no significant differences are found in the primary effect of the learning method, there are significant differences between online learning and the non-adaptive model for hybrid normalisation. Online learning results in a higher loss for both datasets, which is in contrast with our hypothesis H2. The other methods do not show a stable pattern across conditions or datasets.

Most methods perform significantly worse in uniform timing of interaction (random interaction times), as compared to patterned interactions (Gaussian times), supporting our hypothesis H4. Softmax performs equally well on both models for D-All, but performs worse in uniform timing for D-Ten. Hybrid normalisation performs equally well for MMIBN in D-All but performs significantly worse in other conditions.

Hybrid normalisation performs better in all conditions and shows stability across varying conditions compared to the other methods. It achieves lower loss in D-All than in D-Ten, as a result of a higher number of samples in D-All (2,280 in open-set) as compared to D-Ten (800 samples), which shows that the proposed model gets better with the increasing number of recognitions.

Skip C.6Weights and Quality of the Estimation Section

C.6 Weights and Quality of the Estimation

It seems to be self-evident that in the case of uniformly distributed time of interaction, online learning would provide worse results, because the information provided by time will be unreliable. Hence, Bayesian optimisation should find a lower weight for the time parameter. The parameters corresponding to the optimum loss, presented in Figure 19, show otherwise. Weight for the uniform time is higher than that of the Gaussian for online learning in both datasets.

Fig. 19.

Fig. 19. Values of the optimised parameters (the weights of soft biometrics in the MMIBN models and the threshold for the quality of the estimation) through Bayesian optimisation of 303 iterations over 5-fold cross-validation for hybrid normalisation for all samples (D-All, represented with dark colours) and 10 samples (D-Ten, represented with light colours) training sets.

While the average standard deviation of NAOqi age estimation from the true age of the users20 (i.e., the average standard deviation of error) is found to be 11.0 (which was 9.3 in Reference [43]), age is found to be the most important parameter and height the least. This is in contrast with our findings in Reference [43]. However, the results on the Multi-modal Long-term User Recognition Dataset are more generalisable to larger populations, because of the higher number of users (200) and the diverse age range (10–63), in comparison to the limited number of users (14) and the narrow age range (24–40) in our prior work. Note that in the multi-modal dataset, we used the same standard deviation of height estimations (6.3 cm) as Reference [43]. The standard deviation within age estimation (i.e., without using ground truths) is found to be 8.2, which is less than the standard deviation of error. NAOqi gender recognition rate21 is found to be 0.9, and no difference is found between genders, that is, females and males are recognised equally accurately. The optimised threshold for the quality of the estimation (\( \theta _Q \)) is found to be less than 0.1 in each condition. The underlying reason is the disagreement of the modalities, which can decrease the differences in posterior probabilities, because the results are combined through the product rule in the Bayesian network. When the modalities agree with high confidences (probabilities), the quality can be very high, such as \( Q=7.44, \) as shown in Figure 7 in Section 6.5 for the second user.

D EXTREME VALUE MACHINE FOR INCREMENTAL ONLINE LEARNING

EVM22 [73] is a state-of-the-art open world classifier based on the Extreme Value Theory (EVT). However, it was only evaluated using batch learning, which is not suitable in a real-world HRI application, because the users will be encountered sequentially. Hence, we transformed the method for using sequential data and incremental online learning to compare the performance to our proposed methods.23

The hyperparameters of EVM are tail size (\( \tau \), the number of points that constitute extrema for EVT), number of models to average (\( k \)), coverage threshold (\( \varsigma \), probabilistic threshold to designate redundancy between points), and open-set threshold (\( \delta \), if the maximum probability is below this threshold, then the identity is estimated as unknown). The ranges considered for these hyperparameters in Reference [73] are as follows: 100–32,000 for \( \tau \) (can be minimum 2), 1–10 for \( k \), \( [0.008,0.186,0.492,1.0] \) for \( \varsigma \), and \( [0.05,0.1,\ldots ,0.3] \) for \( \delta \). Moreover, Euclidean distance or cosine similarity can be used as the distance function to compute margins for EVM.

As described in Section 3.4, we set MMIBN to declare the user as unknown in the first four recognitions to allow the network to make meaningful estimations. This was achieved for EVM by setting \( \tau =3 \). After the initial training, sequential learning is achieved by updating the model with a single data point (i.e., a single recognition) at each recognition, by setting \( k=1 \). We optimised \( \varsigma \) and \( \delta \) over the ranges given and found that \( \varsigma =1.0 \) and \( \delta =0.05 \) resulted in the lowest long-term recognition performance loss. Cosine similarity is used as the distance function, as it is stated in Reference [73] that Euclidean distance led to poor performance for EVM.

It is important to note that in Reference [73], \( \tau =33998 \), \( k=6 \), and \( \varsigma =0.5 \). However, the authors stated that \( \varsigma \) and \( k \) had a slight impact on performance (2% increase in accuracy and F1 score), whereas the vast majority of performance variation was attributed to \( \tau \).

We use the same data with the structure described in Section 3.1 for evaluating MMIBN and EVM models. Note that for EVM models, the data is not normalised for face recognition similarity scores, and the normal curves for the remaining modalities are normalised through norm-sum (dividing by the total sum), because hybrid normalisation is a feature that we introduced in this article for MMIBN and it is optimised for that structure. Using normalisation for face recognition does not change the performance of EVM:FR, but hybrid normalisation results in a poor performance for EVM:MM (DIR is 0.029).

E TUKEY’S HONESTLY SIGNIFICANT DIFFERENCES TEST PLOTS

In this manuscript, a letter representation is adopted for Tukey’s HSD test plots. Levels that are not significantly different from each other at 0.95 confidence level (\( p\lt .05 \)) are represented with the same letter over all the conditions, that is, each method is compared to all the other methods in different conditions. In other words, if two methods do not share a common letter, then there is a significant difference in performance between them. Multiple letters mean that the method is at the same significance level as multiple other methods.

Skip E.1Long-term Recognition Performance Loss Section

E.1 Long-term Recognition Performance Loss

Figure 20 presents Tukey’s HSD test results on the training, open-set, closed-set (training), closed-set (open) evaluation sets for D-All datasets with Gaussian and uniform timing of interaction. The results show that the proposed MMIBN model significantly outperforms FR and EVM in all of the datasets.

Fig. 20.

Fig. 20. Comparison of Tukey’s HSD test results on loss for the proposed multi-modal incremental Bayesian network (MMIBN), face recognition (FR), soft biometrics (SB) with online learning condition (:OL), Extreme Value Machine with FR data (EVM:FR), and with multi-modal data (EVM:MM). The results are presented for training (100 users), open-set test (200 users), closed-set (training) (100 users) and closed-set (open) (200 users) for all samples dataset (D-All) for Gaussian and uniform timing of interaction. Lower loss is better.

Skip E.2Open-set Identification Metrics: DIR and FAR Section

E.2 Open-set Identification Metrics: DIR and FAR

Tukey’s HSD test results for DIR and FAR are presented in Figures 21 and 22, respectively. The plot for DIR resembles highly that of Figure 20 in a reversed direction, because of \( \alpha *(1-DIR) \) component of loss, whereby, \( \alpha =0.9 \). DIR of MMIBN is significantly higher than FR and EVM in all datasets. The detailed analysis is presented in Section 6.5.2.

Fig. 21.

Fig. 21. Tukey’s HSD test results for detection and identification rate (DIR) of all models for D-All datasets. Higher DIR is better.

Fig. 22.

Fig. 22. Tukey’s HSD test results for false alarm rate (FAR) of all models for D-All datasets. Lower FAR is better.

FAR \( =0 \) in closed-sets, because all the users are previously enrolled. FR has a very low FAR in large datasets, because it predominantly identifies users as unknown. The combination of several modalities increases the probability to mistake a user for another user, which increase FAR in MMIBN.

Skip E.3User-specific Analysis Section

E.3 User-specific Analysis

Figure 23 presents the significant differences between the identification of users within the all samples dataset with patterned times. FR significantly performs better or worse for some of the users, whereas the combination of multi-modalities through our proposed model decreases the bias of FR. Online learning (EVM and MMIBN:OL) further mitigates the user recognition bias, in exchange for the performance, due to the accumulating noise in the identifiers.

Fig. 23.

Fig. 23. Tukey’s HSD test results for significant differences of user-based identification over 5-fold cross-validation on D-AllGaussian: (A) face recognition (FR), (B) proposed model (MMIBN), (C) proposed model with online learning (MMIBN:OL), (D) Extreme Value Machine with face recognition data (EVM:FR). The darker blue colours represent significant differences, whereas lighter blue colours mean that the users are identified equally well.

F TIME PLOT FOR OPEN-SET RECOGNITION

The time plot for open-set recognition in Figure 24 shows the change in long-term recognition loss with the increasing number of recognitions. The results are consistent with the results for the training set, presented in Section 6.5.1. MMIBN and MMIBN:OL have a higher loss in the open-set compared to the training, due to the higher number of users to recognise. EVM:FR has a lower loss during the enrolment period due to lower FAR compared to MMIBN models, and a higher DIR compared to EVM:MM, but the MMIBN models significantly outperform it overall and in the closed-set.

Fig. 24.

Fig. 24. The change of loss with increasing number of recognitions for all samples dataset with Gaussian times (D-AllGaussian) for open-set and closed-set (open). The loss decreases with increasing number of recognitions.

Skip ACKNOWLEDGMENTS Section

ACKNOWLEDGMENTS

The authors would like to thank the reviewers for their valuable suggestions for improving the presentation and analyses of the results, Valerio Biscione for his valuable suggestions in the design of the MMIBN, Pierre-Henri Wuillemin for his substantial help with the pyAgrum library, Ethan Rudd for sharing the Extreme Value Machine code, Jonathan Casas and Nathalia Céspedes Gómez for their help in integrating MMIBN within the personalised patient-robot interface for the cardiac rehabilitation programme, Mehdi Hellou for integrating MMIBN into the barista robot, the participants in our experiments for their time and efforts, and Hoang-Long Cao for his contribution of artwork in the human-robot interaction diagram (Figure 1).

Footnotes

  1. 1 https://www.wired.com/story/10-year-old-face-id-unlocks-mothers-iphone-x/.

    Footnote
  2. 2 Latest version of the Multi-modal Incremental Bayesian Network: https://github.com/birfan/MultimodalRecognition Multi-modal Long-Term User Recognition Dataset, source code used in this work and the corresponding results and the trained models are available at: https://github.com/birfan/MultimodalRecognitionDataset.

    Footnote
  3. 3 https://data.vision.ee.ethz.ch/cvl/rrothe/imdb-wiki/.

    Footnote
  4. 4 https://www.softbankrobotics.com/corp/robots/.

    Footnote
  5. 5 https://agrum.gitlab.io/.

    Footnote
  6. 6 This parameter can be set to another value (including 0) in the algorithm for MMIBN. Increasing this value would allow the MMIBN to produce more reliable estimations of new users, however, this could also decrease the identification performance of known users. Hence, we chose a sufficiently low value. It is also possible to use the identity estimated by face recognition (instead of declaring unknown identity) in the algorithm for the first few recognitions.

    Footnote
  7. 7 https://www.softbankrobotics.com/corp/robots/.

    Footnote
  8. 8 http://doc.aldebaran.com/2-5.

    Footnote
  9. 9 An implementation of Dlib for open world recognition using retraining on a dataset with a small number of users is explained in this link, which shows that a single recognition can take 6–7 seconds on a single CPU system: https://www.pyimagesearch.com/2018/06/18/face-recognition-with-opencv-python-and-deep-learning/.

    Footnote
  10. 10 https://cmusatyalab.github.io/openface.

    Footnote
  11. 11 The classifier demos in https://cmusatyalab.github.io/openface/demo-1-web/ and https://cmusatyalab.github.io/openface/demo-3-classifier/ were combined and applied on the Pepper robot. An image was taken from the robot’s camera, identified with the pre-trained OpenFace celebrity classifier available at the latter link, and the confidence score of the classifier was displayed on the Pepper’s tablet, along with the image of the user and the most similar celebrity. The confidence score of the classification ranges from 0% (user does not resemble any user in the database) to 100% (user is identical to a user in the database). If the confidence score is below (or equal to) 50%, then the user is identified as unknown (new), as defined in the script for the former demo.

    Footnote
  12. 12 https://www.nist.gov/srd/nist-special-database-18.

    Footnote
  13. 13 https://github.com/birfan/MultimodalRecognitionDataset.

    Footnote
  14. 14 https://thuijskens.github.io/2016/12/29/bayesian-optimisation/.

    Footnote
  15. 15 https://github.com/EMRResearch/ExtremeValueMachine.

    Footnote
  16. 16 The results are given for D-All with Gaussian timing on the open-set.

    Footnote
  17. 17 Known user interaction: https://youtu.be/Ix98k6_-2Zc.

    Footnote
  18. 18 Created for MuMMER project: http://mummer-project.eu.

    Footnote
  19. 19 Due to the errors in data, we could not apply statistical comparison between the MMIBN models and the baselines.

    Footnote
  20. 20 The standard deviation of age estimation from the ground truth values are calculated per user, averaged over 200 users, and then averaged over 5-folds within the all samples dataset with Gaussian times (D-AllGaussian).

  21. 21 The gender recognition rate is the fraction of correctly estimated gender in the images (based on ground truths) of 200 users, averaged over 5-folds within the all samples dataset with Gaussian times (D-AllGaussian).

  22. 22 https://github.com/EMRResearch/ExtremeValueMachine.

  23. 23 Modified version of the Extreme Value Machine is provided online: https://github.com/birfan/MultimodalRecognitionDataset.

REFERENCES

  1. [1] Al-Qaderi Mohammad K. and Rad Ahmad B.. 2018. A multi-modal person recognition system for social robots. Appl. Sci. 8, 3 (2018). DOI: https://doi.org/10.3390/app8030387Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Amos Brandon, Ludwiczuk Bartosz, and Satyanarayanan Mahadev. 2016. OpenFace: A General-purpose Face Recognition Library with Mobile Applications. Technical Report. CMU-CS-16-118, CMU School of Computer Science.Google ScholarGoogle Scholar
  3. [3] Andry Pierre, Gaussier Philippe, Moga Sorin, Banquet Jean Paul, and Nadel Jacqueline. 2001. Learning and communication via imitation: An autonomous robot perspective. IEEE Trans. Syst., Man, Cyber. - Part A: Syst. Hum. 31, 5 (2001), 431442. DOI: https://doi.org/10.1109/3468.952717Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. [4] Aryananda Lijin. 2001. Online and unsupervised face recognition for humanoid robot: Toward relationship with people. In IEEE-RAS International Conference on Humanoid Robots. IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Aryananda Lijin. 2009. Learning to recognize familiar faces in the real world. In IEEE International Conference on Robotics and Automation (ICRA). IEEE, 19911996. DOI: https://doi.org/10.1109/ROBOT.2009.5152362Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Bauer Eric, Koller Daphne, and Singer Yoram. 1997. Update rules for parameter estimation in Bayesian networks. In 13th Conference on Uncertainty in Artificial Intelligence. 313.Google ScholarGoogle Scholar
  7. [7] Beleites Claudia and Salzer Reiner. 2008. Assessing and improving the stability of chemometric models in small sample size situations. Analyt. Bioanalyt. Chem. 390, 5 (2008), 12611271.Google ScholarGoogle ScholarCross RefCross Ref
  8. [8] Bendale Abhijit and Boult Terrance. 2015. Towards open world recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 18931902. DOI: https://doi.org/10.1109/CVPR.2015.7298799Google ScholarGoogle ScholarCross RefCross Ref
  9. [9] Bendale Abhijit and Boult Terrance E.. 2016. Towards open set deep networks. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 15631572. DOI: https://doi.org/10.1109/CVPR.2016.173Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Bigün E. S., Bigün J., Duc B., and Fischer S.. 1997. Expert conciliation for multi modal person authentication systems by Bayesian statistics. In Audio- and Video-based Biometric Person Authentication. (Lecture Notes in Computer Science, Vol. 1206.) Springer, Berlin. DOI: https://doi.org/10.1007/BFb0016008Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Bishop Christopher M.. 2006. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer-Verlag New York, Inc.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. [12] Boucenna Sofiane, Cohen David, Meltzoff Andrew N., Gaussier Philippe, and Chetouani Mohamed. 2016. Robots learn to recognize individuals from imitative encounters with people and avatars. Sci. Rep. 6 (2016), 19908. DOI: https://doi.org/10.1038/srep19908Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Boucenna Sofiane, Gaussier Philippe, Andry Pierre, and Hafemeister Laurence. 2014. A robot learns the facial expressions recognition and face/non-face discrimination through an imitation game. Int. J. Soc. Robot. 6 (2014), 633652. DOI: https://doi.org/10.1007/s12369-014-0245-zGoogle ScholarGoogle ScholarCross RefCross Ref
  14. [14] Casas Jonathan, Gomez Nathalia Céspedes, Senft Emmanuel, Irfan Bahar, Gutiérrez Luisa F., Rincón Monica, Múnera Marcela, Belpaeme Tony, and Cifuentes Carlos A.. 2018. Architecture for a social assistive robot in cardiac rehabilitation. In IEEE 2nd Colombian Conference on Robotics and Automation (CCRA). 16. DOI: https://doi.org/10.1109/CCRA.2018.8588133Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Céspedes Natalia, Irfan Bahar, Senft Emmanuel, Cifuentes Carlos A., Gutierrez Luisa F., Rincon-Roncancio Mónica, Belpaeme Tony, and Múnera Marcela. 2021. A socially assistive robot for long-term cardiac rehabilitation in the real world. Front. Neurorobot. 15 (2021), 21. DOI: https://doi.org/10.3389/fnbot.2021.633248Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Chen Ching-Han and Chu Chia Te. 2005. Fusion of face and iris features for multimodal biometrics. In Advances in Biometrics. Springer, 571580. DOI: https://doi.org/10.1007/11608288_76Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. [17] Choudhury Tanzeem, Clarkson Brian, Jebara Tony, and Pentland Alex. 1999. Multimodal person recognition using unconstrained audio and video. In International Conference on Audio and Video-based Person Authentication. 176181.Google ScholarGoogle Scholar
  18. [18] Churamani Nikhil, Anton Paul, Brügger Marc, Flieundefinedwasser Erik, Hummel Thomas, Mayer Julius, Mustafa Waleed, Ng Hwei Geok, Nguyen Thi Linh Chi, Nguyen Quan, al. et2017. The impact of personalisation on human-robot interaction in learning scenarios. In 5th International Conference on Human Agent Interaction. Association for Computing Machinery, 171180. DOI: https://doi.org/10.1145/3125739.3125756Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. [19] Clabaugh Caitlyn, Mahajan Kartik, Jain Shomik, Pakkar Roxanna, Becerra David, Shi Zhonghao, Deng Eric, Lee Rhianna, Ragusa Gisele, and Matarić Maja. 2019. Long-term personalization of an in-home socially assistive robot for children with autism spectrum disorders. Front. Robot. AI 6 (2019), 110. DOI: https://doi.org/10.3389/frobt.2019.00110Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Cohen Ira, Bronstein Alexandre, and Cozman Fabio G.. 2001. Online Learning of Bayesian Network Parameters. Technical Report HPL-2001-55 (R.1). HP Laboratories.Google ScholarGoogle Scholar
  21. [21] Connaughton Ryan, Bowyer Kevin W., and Flynn Patrick J.. 2016. Fusion of face and iris biometrics. In Handbook of Iris Recognition. Springer London, 397415. DOI: https://doi.org/10.1007/978-1-4471-6784-6_17Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Cooper Gregory F. and Herskovits Edward. 1992. A Bayesian method for the induction of probabilistic networks from data. Mach. Learn. 9 (1992), 309347. DOI: https://doi.org/10.1007/BF00994110Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Cruz Claudia, Sucar L. Enrique, and Morales Eduardo F.. 2008. Real-time face recognition for human-robot interaction. In 8th IEEE International Conference on Automatic Face Gesture Recognition. IEEE, 16.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Dantcheva Antitza, Elia Petros, and Ross Arun. 2016. What else does your biometric data reveal? A survey on soft biometrics. IEEE Trans. Inf. Forens. Secur. 11, 3 (2016), 441467. DOI: https://doi.org/10.1109/TIFS.2015.2480381Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. [25] Dantcheva Antitza, Velardo Carmelo, D’Angelo Angela, and Dugelay Jean-Luc. 2011. Bag of soft biometrics for person identification. Multimed. Tools Appl. 51 (2011), 739777. DOI: https://doi.org/10.1007/s11042-010-0635-7Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Rosa Rocco De, Mensink Thomas, and Caputo Barbara. 2016. Online Open World Recognition. arxiv:cs.CV/1604.02275Google ScholarGoogle Scholar
  27. [27] Rosa Rocco D. De, Orabona Francesco, and Cesa-Bianchi Nicolò. 2015. The ABACOC algorithm: A novel approach for nonparametric classification of data streams. In IEEE International Conference on Data Mining. IEEE, 733738. DOI: https://doi.org/10.1109/ICDM.2015.43Google ScholarGoogle Scholar
  28. [28] Dietterich Thomas G.. 1998. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 10, 7 (1998), 18951923. DOI: https://doi.org/10.1162/089976698300017197Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. [29] Fei Geli, Wang Shuai, and Liu Bing. 2016. Learning cumulatively to become more knowledgeable. In 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD’16). ACM, 15651574. DOI: https://doi.org/10.1145/2939672.2939835Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. [30] Feil-Seifer David and Matarić Maja J.. 2005. Defining socially assistive robotics. In 9th International Conference on Rehabilitation Robotics (ICORR). 465468. DOI: https://doi.org/10.1109/ICORR.2005.1501143Google ScholarGoogle Scholar
  31. [31] Filliat David. 2007. A visual bag of words method for interactive qualitative localization and mapping. In IEEE International Conference on Robotics and Automation. IEEE, 39213926. DOI: https://doi.org/10.1109/ROBOT.2007.364080Google ScholarGoogle ScholarCross RefCross Ref
  32. [32] Gaisser Floris, Rudinac Maja, Jonker Pieter P., and Tax David. 2013. Online face recognition and learning for cognitive robots. In 16th International Conference on Advanced Robotics (ICAR). IEEE, 19. DOI: https://doi.org/10.1109/ICAR.2013.6766498Google ScholarGoogle ScholarCross RefCross Ref
  33. [33] Ge ZongYuan, Demyanov Sergey, Chen Zetao, and Garnavi Rahil. 2017. Generative OpenMax for Multi-Class Open Set Classification. arxiv:cs.CV/1707.07418Google ScholarGoogle Scholar
  34. [34] Gepperth A. and Hammer B.. 2016. Incremental learning algorithms and applications. In European Symposium on Artificial Neural Networks (ESANN).Google ScholarGoogle Scholar
  35. [35] Giuliano Catherine, Parmenter Belinda J., Baker Michael K., Mitchell Braden L., Williams Andrew D., Lyndon Katie, Mair Tarryn, Maiorana Andrew, Smart Neil A., and Levinger Itamar. 2017. Cardiac rehabilitation for patients with coronary artery disease: A practical guide to enhance patient outcomes through continuity of care. Clin. Med. Insights Cardiol. 11 (2017). DOI: https://doi.org/10.1177/1179546817710028Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Gonzales Christophe, Torti Lionel, and Wuillemin Pierre-Henri. 2017. aGrUM: A graphical universal model framework. In International Conference on Industrial Engineering, Other Applications of Applied Intelligent Systems. Springer. DOI: https://doi.org/10.1007/978-3-319-60045-1_20Google ScholarGoogle ScholarCross RefCross Ref
  37. [37] Hampel Frank R., Ronchetti Elvezio M., Rousseeuw Peter J., and Stahel Werner A.. 1986. Robust Statistics: The Approach Based on Influence Functions. Wiley.Google ScholarGoogle Scholar
  38. [38] Hanheide Marc, Wrede Sebastian, Lang Christian, and Sagerer Gerhard. 2008. Who am I talking with? A face memory for social robots. In IEEE International Conference on Robotics and Automation, ICRA. IEEE, 36603665. DOI: https://doi.org/10.1109/ROBOT.2008.4543772Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Heckerman David, Geiger Dan, and Chickering David M.. 1995. Learning Bayesian networks: The combination of knowledge and statistical data. Mach. Learn. 20 (1995), 197243. DOI: https://doi.org/10.1023/A:1022623210503Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Hue Carine, Boullé Marc, and Lemaire Vincent. 2017. Online Learning of a Weighted Selective Naive Bayes Classifier with Non-convex Optimization. Springer International Publishing, Cham, 317. DOI: https://doi.org/10.1007/978-3-319-45763-5_1Google ScholarGoogle Scholar
  41. [41] Irfan Bahar, Gomez Natalia Céspedes, Casas Jonathan, Senft Emmanuel, Gutiérrez Luisa F., Rincon-Roncancio Monica, Munera Marcela, Belpaeme Tony, and Cifuentes Carlos A.. 2020. Using a personalised socially assistive robot for cardiac rehabilitation: A long-term case study. In 29th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). IEEE, 124130. DOI: https://doi.org/10.1109/RO-MAN47096.2020.9223491Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Irfan Bahar, Hellou Mehdi, Mazel Alexandre, and Belpaeme Tony. 2020. Challenges of a real-world HRI study with non-native English speakers: Can personalisation save the day? In 15th ACM/IEEE International Conference on Human-robot Interaction (HRI). ACM. DOI: https://doi.org/10.1145/3371382.3378278Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Irfan Bahar, Lyubova Natalia, Ortiz Michael Garcia, and Belpaeme Tony. 2018. Multi-modal open-set person identification in HRI. In ACM/IEEE International Conference on Human-robot Interaction Social Robots in the Wild Workshop. ACM.Google ScholarGoogle Scholar
  44. [44] Jain Anil, Nandakumar Karthik, and Ross Arun. 2005. Score normalization in multimodal biometric systems. Pattern Recog. 38, 12 (2005), 22702285. DOI: https://doi.org/10.1016/j.patcog.2005.01.012Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. [45] Jain Anil K., Dass Sarat C., and Nandakumar Karthik. 2004. Soft biometric traits for personal recognition systems. In International Conference on Biometric Authentication (LNCS). Springer, Hong Kong, China, 731738. DOI: https://doi.org/10.1007/978-3-540-25948-0_99Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Jain Anil K. and Park Unsang. 2009. Facial marks: Soft biometric for face recognition. In 16th IEEE International Conference on Image Processing (ICIP). IEEE Press, 3740. DOI: https://doi.org/10.1109/ICIP.2009.5413921Google ScholarGoogle ScholarCross RefCross Ref
  47. [47] Jain Anil K., Ross Arun A., and Nandakumar Karthik. 2011. Multibiometrics. In Introduction to Biometrics. Springer Publishing Company, Incorporated, 209258.Google ScholarGoogle ScholarCross RefCross Ref
  48. [48] Jain Lalit P., Scheirer Walter J., and Boult Terrance E.. 2014. Multi-class open set recognition using probability of inclusion. In Computer Vision – ECCV 2014, Fleet David, Pajdla Tomas, Schiele Bernt, and Tuytelaars Tinne (Eds.). Springer International Publishing, Zurich, Switzerland, 393409.Google ScholarGoogle ScholarCross RefCross Ref
  49. [49] Kanungo Tapas, Mount David M., Netanyahu Nathan S., Piatko Christine D., Silverman Ruth, and Wu Angela Y.. 2002. An efficient k-means clustering algorithm: analysis and implementation. IEEE Trans. Pattern Anal. Mach. Intell. 24, 7 (2002), 881892. DOI: https://doi.org/10.1109/TPAMI.2002.1017616Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. [50] King Davis E.. 2009. Dlib-ml: A machine learning toolkit. J. Mach. Learn. Res. 10 (2009), 17551758.Google ScholarGoogle ScholarDigital LibraryDigital Library
  51. [51] Koller Daphne and Friedman Nir. 2009. Parameter estimation. In Probabilistic Graphical Models: Principles and Techniques. The MIT Press, 717782.Google ScholarGoogle Scholar
  52. [52] Kraus William E. and Keteyian Steven J.. 2007. Cardiac Rehabilitation. Humana Press, Totowa, NJ. DOI: https://doi.org/10.1007/978-1-59745-452-0Google ScholarGoogle ScholarCross RefCross Ref
  53. [53] Lambert Alexis, Norouzi Nahal, Bruder Gerd, and Welch Gregory. 2020. A systematic review of ten years of research on human interaction with social robots. Int. J. Hum.comput. Interact. 36, 19 (2020), 18041817. DOI: https://doi.org/10.1080/10447318.2020.1801172Google ScholarGoogle ScholarCross RefCross Ref
  54. [54] Lara Juan S., Casas Jonathan, Aguirre Andres, Munera Marcela, Rincon-Roncancio Monica, Irfan Bahar, Senft Emmanuel, Belpaeme Tony, and Cifuentes Carlos A.. 2017. Human-robot sensor interface for cardiac rehabilitation. In International Conference on Rehabilitation Robotics (ICORR). IEEE, 10131018. DOI: https://doi.org/10.1109/ICORR.2017.8009382Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. [55] Lee Kuang-Chih and Kriegman D.. 2005. Online learning of probabilistic appearance manifolds for video-based recognition and tracking. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), Vol. 1. IEEE, 852859. DOI: https://doi.org/10.1109/CVPR.2005.260Google ScholarGoogle Scholar
  56. [56] Li Ming and Zhang Ren. 2020. Evolving a weighted Bayesian network for consequence assessment of terrorist attack. IEEE Access 8 (2020), 8828288293. DOI: https://doi.org/10.1109/ACCESS.2020.2993016Google ScholarGoogle ScholarCross RefCross Ref
  57. [57] Lim Sungsoo and Cho Sung-Bae. 2006. Online learning of Bayesian network parameters with incomplete data. In Computational Intelligence, Huang De-Shuang, Li Kang, and Irwin George William (Eds.). Springer Berlin, 309314.Google ScholarGoogle ScholarCross RefCross Ref
  58. [58] Lin Chao-Yu, Song Kai-Tai, Chen Yi-Wen, Chien Shuo-Cheng, Chen Sin-Horng, Chiang Chen-Yu, Yang Jyh-Her, Wu Yi-Chiao, and Liu Tzu-Jui. 2012. User identification design by fusion of face recognition and speaker recognition. In 12th International Conference on Control, Automation and Systems. 14801485.Google ScholarGoogle Scholar
  59. [59] Liu Jinzhong and Liao Qin. 2008. Online learning of Bayesian network parameters. In 4th International Conference on Natural Computation, Vol. 3. IEEE, 267271. DOI: https://doi.org/10.1109/ICNC.2008.651Google ScholarGoogle ScholarDigital LibraryDigital Library
  60. [60] Mangiafico Salvatore S.. 2016. Summary and Analysis of Extension Program Evaluation in R. Vol. version 1.18.8. Retrieved from: rcompanion.org/handbook/.Google ScholarGoogle Scholar
  61. [61] Martinson Eric, Lawson Wallace, and Trafton Greg. 2013. Identifying people with soft-biometrics at fleet week. In 8th ACM/IEEE International Conference on Human-robot Interaction (HRI’13). IEEE Press, 4956.Google ScholarGoogle ScholarDigital LibraryDigital Library
  62. [62] McClelland James L., McNaughton Bruce L., and O’Reilly Randall C.. 1995. Why there are complementary learning systems in the hippocampus and neocortex: Insights from the successes and failures of connectionist models of learning and memory. Psychol. Rev. 102, 3 (1995), 419457.Google ScholarGoogle ScholarCross RefCross Ref
  63. [63] McCloskey Michael and Cohen Neal J.. 1989. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of Learning and Motivation, Bower Gordon H. (Ed.). Vol. 24. Academic Press, 109165. DOI: https://doi.org/10.1016/S0079-7421(08)60536-8Google ScholarGoogle Scholar
  64. [64] Mensink Thomas, Verbeek Jakob, Perronnin Florent, and Csurka Gabriela. 2013. Distance-based image classification: Generalizing to new classes at near-zero cost. IEEE Trans. Pattern Anal. Mach. Intell. 35, 11 (2013), 26242637. DOI: https://doi.org/10.1109/TPAMI.2013.83Google ScholarGoogle ScholarDigital LibraryDigital Library
  65. [65] Ouellet Simon, Grondin François, Leconte Francis, and Michaud François. 2014. Multimodal biometric identification system for mobile robots combining human metrology to face recognition and speaker identification. In 23rd IEEE International Symposium on Robot and Human Interactive Communication. IEEE, 323328. DOI: https://doi.org/10.1109/ROMAN.2014.6926273Google ScholarGoogle ScholarCross RefCross Ref
  66. [66] Parisi German Ignacio, Kemker Ronald, Part Jose L., Kanan Christopher, and Wermter Stefan. 2018. Continual Lifelong Learning with Neural Networks: A Review. arxiv:cs.LG/1802.07569Google ScholarGoogle Scholar
  67. [67] Park Unsang and Jain Anil K.. 2010. Face matching and retrieval using soft biometrics. IEEE Trans. Inf. Forens. Secur. 5, 3 (Sept. 2010), 406415. DOI: https://doi.org/10.1109/TIFS.2010.2049842Google ScholarGoogle ScholarDigital LibraryDigital Library
  68. [68] Parkhi Omkar M., Vedaldi Andrea, and Zisserman Andrew. 2015. Deep face recognition. In British Machine Vision Conference, Xie Xianghua, Jones Mark W., and Tam Gary K. L. (Eds.). BMVA Press, Swansea, UK, 41.1–41.12. DOI: https://doi.org/10.5244/C.29.41Google ScholarGoogle ScholarCross RefCross Ref
  69. [69] Phillips P. Jonathon, Grother Patrick, and Micheals Ross. 2011. Evaluation methods in face recognition. In Handbook of Face Recognition (2nd ed.), Li Stan Z. and Jain Anil K. (Eds.). Springer Publishing Company, Incorporated, 553556.Google ScholarGoogle ScholarCross RefCross Ref
  70. [70] Richardson Kathleen, Coeckelbergh Mark, Wakunuma Kutoma, Billing Erik, Ziemke Tom, Gomez Pablo, Vanderborght Bram, and Belpaeme Tony. 2018. Robot enhanced therapy for children with autism (DREAM): A social model of autism. IEEE Technol. Soc. Mag. 37, 1 (2018), 3039.Google ScholarGoogle ScholarCross RefCross Ref
  71. [71] Rothe Rasmus, Timofte Radu, and Gool Luc Van. 2015. DEX: Deep expectation of apparent age from a single image. In IEEE International Conference on Computer Vision Workshops (ICCVW).Google ScholarGoogle ScholarDigital LibraryDigital Library
  72. [72] Rothe Rasmus, Timofte Radu, and Gool Luc Van. 2018. Deep expectation of real and apparent age from a single image without facial landmarks. Int. J. Comput. Vis. 126, 2–4 (2018), 144157.Google ScholarGoogle ScholarDigital LibraryDigital Library
  73. [73] Rudd Ethan M., Jain Lalit P., Scheirer Walter J., and Boult Terrance E.. 2018. The extreme value machine. IEEE Trans. Pattern Anal. Mach. Intell. 40, 3 (2018), 762768.Google ScholarGoogle ScholarCross RefCross Ref
  74. [74] Sadhya Debanjan, Pahariya Parth, Yadav Rishi, Rastogi Apoorv, Kumar Ayush, Sharma Lakshya, and Singh Sanjay K.. 2017. BioSoft—A multimodal biometric database incorporating soft traits. In IEEE International Conference on Identity, Security and Behavior Analysis (ISBA). IEEE, 16. DOI: https://doi.org/10.1109/ISBA.2017.7947693Google ScholarGoogle ScholarCross RefCross Ref
  75. [75] Scassellati Brian, Boccanfuso Laura, Huang Chien-Ming, Mademtzi Marilena, Qin Meiying, Salomons Nicole, Ventola Pamela, and Shic Frederick. 2018. Improving social skills in children with ASD using a long-term, in-home social robot. Sci. Robot. 3, 21 (2018). DOI: https://doi.org/10.1126/scirobotics.aat7544Google ScholarGoogle ScholarCross RefCross Ref
  76. [76] Scheirer Walter J., Rocha Anderson de Rezende, Sapkota Archana, and Boult Terrance E.. 2013. Toward open set recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35, 7 (2013), 17571772. DOI: https://doi.org/10.1109/TPAMI.2012.256Google ScholarGoogle ScholarDigital LibraryDigital Library
  77. [77] Scheirer Walter J., Jain Lalit P., and Boult Terrance E.. bibinfoyear2014. Probability models for open set recognition. IEEE Trans. Pattern Anal. Mach. Intell. 36, 11 (2014), 23172324. DOI: https://doi.org/10.1109/TPAMI.2014.2321392Google ScholarGoogle ScholarCross RefCross Ref
  78. [78] Scheirer W. J., Kumar N., Ricanek K., Belhumeur P. N., and Boult T. E.. 2011. Fusing with context: A Bayesian approach to combining descriptive attributes. In International Joint Conference on Biometrics (IJCB). IEEE, 18. DOI: https://doi.org/10.1109/IJCB.2011.6117490Google ScholarGoogle ScholarDigital LibraryDigital Library
  79. [79] Schroff Florian, Kalenichenko Dmitry, and Philbin James. 2015. FaceNet: A unified embedding for face recognition and clustering. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). DOI: https://doi.org/10.1109/cvpr.2015.7298682Google ScholarGoogle ScholarCross RefCross Ref
  80. [80] Sun Yi, Wang Xiaogang, and Tang Xiaoou. 2014. Deep learning face representation from predicting 10,000 classes. In IEEE Conference on Computer Vision and Pattern Recognition. IEEE, 18911898. DOI: https://doi.org/10.1109/CVPR.2014.244Google ScholarGoogle ScholarDigital LibraryDigital Library
  81. [81] Taigman Yaniv, Yang Ming, Ranzato Marc’Aurelio, and Wolf Lior. 2014. DeepFace: Closing the gap to human-level performance in face verification. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR’14). IEEE Computer Society, 17011708. DOI: https://doi.org/10.1109/CVPR.2014.220Google ScholarGoogle ScholarDigital LibraryDigital Library
  82. [82] Verlinde Patrick, Druyts Pascal, Cholet Gerard, and Acheroy Marc. 1999. Applying Bayes based classifiers for decision fusion in a multi-modal identity verification system. In International Symposium on Pattern Recognition.Google ScholarGoogle Scholar
  83. [83] Wang Yunhong, Tan Tieniu, and Jain Anil K.. 2003. Combining face and iris biometrics for identity verification. In 4th International Conference on Audio- and Video-based Biometric Person Authentication (AVBPA’03). Springer-Verlag, Berlin, 805813.Google ScholarGoogle ScholarDigital LibraryDigital Library
  84. [84] Winkle Katie, Caleb-Solly Praminda, Turton Ailie, and Bremner Paul. 2018. Social robots for engagement in rehabilitative therapies: Design implications from a study with therapists. In ACM/IEEE International Conference on Human-robot Interaction (HRI’18). Association for Computing Machinery, New York, NY, 289297. DOI: https://doi.org/10.1145/3171221.3171273Google ScholarGoogle ScholarDigital LibraryDigital Library
  85. [85] Wójcik Waldemar, Gromaszek Konrad, and Junisbekov Muhtar. 2016. Face recognition: Issues, methods and alternative applications. In Face Recognition—Semisupervised Classification, Subspace Projection and Evaluation Methods, Ramakrishnan S. (Ed.). InTech. DOI: https://doi.org/10.5772/61471Google ScholarGoogle ScholarCross RefCross Ref
  86. [86] Zewail Rami, Elsafi Ahmed, Saeb Magdy, and Hamdy Nourhan. 2004. Soft and hard biometrics fusion for improved identity verification. In 47th Midwest Symposium on Circuits and Systems, Vol. 1. IEEEI–225. DOI: https://doi.org/10.1109/MWSCAS.2004.1353967Google ScholarGoogle ScholarCross RefCross Ref
  87. [87] Zhang Zhijian, Wang Rui, Pan Ke, Li Stan Z., and Zhang Peiren. 2007. Fusion of near infrared face and iris biometrics. In Advances in Biometrics. Springer Berlin, 172180. DOI: https://doi.org/10.1007/978-3-540-74549-5_19Google ScholarGoogle ScholarCross RefCross Ref
  88. [88] Zhou Yue and Huang T. S.. 2006. Weighted Bayesian network for visual tracking. In 18th International Conference on Pattern Recognition (ICPR’06), Vol. 1. IEEE, 523526. DOI: https://doi.org/10.1109/ICPR.2006.1188Google ScholarGoogle Scholar

Index Terms

  1. Multi-modal Open World User Identification

            Recommendations

            Comments

            Login options

            Check if you have access through your login credentials or your institution to get full access on this article.

            Sign in

            Full Access

            PDF Format

            View or Download as a PDF file.

            PDF

            eReader

            View online with eReader.

            eReader

            HTML Format

            View this article in HTML Format .

            View HTML Format