Professional Documents
Culture Documents
I. INTRODUCTION
Full-body tactile sensitive robotic skins based on large
pressure sensitive arrays has found several applications in
manipulation, exploration, navigation, and human-robotic in-
teraction [1]. Tactile sensing modalities have been especially
useful for conveying emotion and intent in therapeutic and Fig. 1. Tactile sensitive companion robot interacting through affective
companion robots [2]. Over the last few years, detecting touch (from [3]) and basis for the HAART dataset.
and identifying high-level affective touch gestures have been
explored for social human-robot interaction [3][5]. Specif-
ically, there have been several robotic pets developed for learning approaches for social touch recognition has not
therapeutic and companion purposes for the elderly and ill, been explored, with the exception of using deep autoen-
such as in Figure 1. coders to compress individual frames in [10]. Deep learn-
Motivated by the need for high quality gesture classiers, ing approaches have shown several advantages over hand-
two affective touch datasets [4], [6] have been made publicly engineered features: CNNs are useful for automatically gen-
available, and has allowed for several different classication erating features from training data, features exhibit invariance
approaches to be explored [7][10]. These investigations to translation and scaling, and RNNs can provide predictions
have focused primarily on manually developing features at each point in a sequence, rather than one prediction for a
to be extracted from gestures, usually based on common sequence as a whole.
features from other domains, such as speech recognition In addition, we are interested in framing pressure-sensitive
and image processing. In general, the features used in these robotic skins as a robotic material [11], [12]. In the robotic
investigations are extracted from a gesture as a whole, which material paradigm, a network of inexpensive sensing and
requires detecting the onset and end of a gesture, and limits computing elements are embedded in and tightly coupled
the ability to perform real-time classication. with a physical material. Individual computing elements pro-
We are interested in using deep learning techniques, cess local sensor data and share information with neighboring
such as convolutional (CNN) and recurrent neural networks nodes in order to perform some global tasks, such as gesture
(RNN), for classifying social touch gestures in the context classication. This effectively allows a signicant amount of
of robotic materials. To the best of our knowledge, deep computing to be performed in the material itself, reducing
communication to and processing in a centralized computer,
D. Hughes was supported by the Airforce Ofce of Scientic Research.
We are grateful for this support. and allowing the material to scale.
1 Department of Computer Science, University of The purpose of this investigation is to explore the suitabil-
Colorado Boulder, Boulder, CO 80309-0430, USA ity of deep learning for affective touch classication, deter-
{dana.hughes,nikolaus.correll}@colorado.edu
2 The Aerospace Corporation, El Segundo, CA 90245 mine the accuracy trade-off when classifying entire gestures
alon.krauthammer@aero.org versus providing predictions several times per second, and
2316
classication layer. The convolution layers implement 3D D. Autoencoder-RNN
convolutional kernels, described by ve values (time, frame The nal architecture extends the autoencoder used in [10]
width, frame height, number of input channels, number of by connecting the output of the code layer to an RNN layer,
output channels), and the max-pool layers have a pool size as shown in Figure 3. In [10], an autoencoder [15] was used
and stride of 2 in each dimension. The output of the max- to reduce the size of individual frames from 64 units to
pooling layers and fully connected layer are activated with 10 linear units. This output was used as observations for
a rectied linear unit (ReLU) activiation function. hidden Markov models (HMM) to estimate the likelihood
To determine the hyperparameters (i.e., kernel dimen- that a sequence of frames belonged to each type of ges-
sions), multiple CNNs were trained to classify individual ture. In this investigation, the HMM was replaced with a
windows. To determine the number of and dimensionality recurrent layer. The resulting architecture has the advantage
of kernels for each layer, a coarse grid search over these of allowing the autoencoder portion to be pre-trained to
parameters were performed iteratively for two layers. The reconstruct individual frames (as in [10]) prior to training the
kernel parameters were selected by attening the output of classier. Additionally, training the network for classication
the CNN layer being optimized, and connecting this output renes the encoding portion of the network so that the
to a softmax layer, ignoring the intermediate fully connected encoded frame provides better information for classication,
layer. The network was trained to convergence, and the as opposed to simply performing data compression.
kernel parameters were selected from the model that resulted The autoencoder portion of the architecture used matches
in the lowest validation cost without showing evidence of that in [10]: the encoding layers consist of four fully con-
overtting of the training data. nected layers with 200, 100, 50 and 25 units, followed by
The resulting CNN for the CoST dataset consisted a code layer of 10 units. The decoding layer is simply a
of two layersthe rst layers kernel dimensions were mirror of the encoding layer: 25, 50 100, and 200 units. In
(20, 2, 2, 1, 30), and the second layers kernel dimensions [10], the code layer used a linear activation function, while
were (5, 2, 2, 30, 10). The CNN for the HAART dataset con- the encoding and decoding layers consisted of sigmoid units.
sisted of two layersthe rst layers kernel dimensions were Here, rectied linear units were used instead of sigmoid
(10, 3, 3, 1, 30), and the second layers kernel dimensions units, as these types of units learn faster than sigmoid units.
were (5, 2, 2, 30, 20). The number of units in the RNN layer and fully connected
The two convolutional / max-pooling layers were fully output layer was determined similarly to the CNN-RNN
connected to a layer of 50 ReLU units, which is then architecture. Both the RNN and fully connected layers used
connected to the softmax layer for nal classication pur- 25 units; a larger number of units typically resulted in a
poses. The number of units were selected by determining the model which would output a single class regardless of input,
number of units at which validation cost stopped improving. and would show little or no change in validation cost during
training.
2317
Fig. 2. Convolutional Neural Network Architecture. Filter dimensions for each layer and window duration are arbitrarily selected.
TABLE I
S UMMARY OF RESULTS FOR THE THREE MODELS AND [5].
Test
F1 Accuracy
CNN 0.4125 42.34%
CNN-RNN 0.4724 52.86%
Autoencoder-RNN 0.3253 33.52%
CoST
Balli Altuglu et al. [7] 26.00%
Gaus et al. [8] 58.70%
Hughes et al. [10] 47.20%
Ta et al. [9] 61.30%
CNN 0.5538 56.10%
CNN-RNN 0.6028 61.35%
HAART
2318
TABLE III
RNN, which resulted in poor classication accuracy, demon-
C ONFUSION M ATRIX FOR THE HAART DATASET USING CNN-RNN
strates a bias towards the Tickle gesture, though the cause of
ARCHITECTURE .
this is unknown.
Predicted Class
B. Model Size True Class A B C D E F G
(A) No Touch 97.2 0.0 0.0 0.0 0.0 2.8 0.0
The second aspect of interest is the overall memory (B) Constant 2.9 85.7 0.0 5.7 2.9 2.9 0.0
footprint of each model. Models which can be implemented (C) Pat 2.8 0.0 41.7 13.9 16.7 5.6 19.4
on small, inexpensive microcontrollers allow for ofoading (D) Rub 0.0 16.7 11.1 27.8 27.8 13.9 2.8
(E) Scratch 0.0 0.0 5.6 5.6 77.8 0.0 11.1
the task of classication onto the material itself, allowing for (F) Stroke 2.8 5.6 8.3 16.7 2.8 61.1 2.8
only the predicted gesture to be communicated to an external (G) Tickle 0.0 0.0 5.6 0.0 52.8 2.8 38.9
computing sink. Assuming that models are trained off-line on
a larger (e.g., desktop) computer, memory requirements can TABLE IV
be divided into model parameters (i.e., weights and biases) C ONFUSION M ATRIX FOR THE HAART DATASET USING
and neuron activation levels. Measurements and neuron ac- AUTOENCODER -RNN ARCHITECTURE .
tivation levels need to be stored in RAM, which is typically
very limited on microcontrollers (e.g., 8kB on an XMega128 Predicted Class
series), while model parameters can be stored in program True Class A B C D E F G
memory (FLASH), which is generally more plentiful, though (A) No Touch 91.7 0.0 0.0 2.8 0.0 5.6 0.0
still limited (e.g., 256kB on an XMega128). (B) Constant 2.9 97.1 0.0 0.0 0.0 0.0 0.0
(C) Pat 2.8 0.0 47.2 8.3 19.4 19.4 2.8
The memory requirements can be determined directly from (D) Rub 0.0 5.6 0.0 5.6 77.8 0.0 11.1
each model: the number of values stored in RAM is the (E) Scratch 0.0 0.0 11.1 2.8 86.1 0.0 0.0
(F) Stroke 16.7 0.0 8.3 5.6 13.9 44.4 11.1
number of data points per sample window and total number (G) Tickle 0.0 0.0 13.9 0.0 75.0 0.0 11.1
of neurons in the model, and the number of values stored
in FLASH is the number of weights and biases in the
model. We summarize these two values for each model in
Table VIII. For the CNN and CNN-RNN models, convolution The total memory requirements is dependent on the num-
and pooling can be performed jointly, reducing the need to ber of bits required to represent a single value. From the
store the full output of each convolutional layer. We report values given in Table VIII, the maximum amount of FLASH
the number of sensors and neurons for both implementations, required (Autoencoder-RNN model) would be 80kB or
including full convolutional outputs in parentheses. 160kB for 16-bit and 32-bit values, respectively, and the
For the two sets of models, the Autoencoder-RNN requires maximum amount of RAM required (CNN-RNN model)
the largest amount of FLASH memory, primarily due to the would be 13kB (16-bit) or 26kB (32-bit). Based on these
large number of full connections. In contrast, the convolu- values, a 16-bit implementation of any of these models could
tional kernels in the CNN and CNN-RNN models provide a be implemented on an inexpensive ($5) microcontroller.
large amount of weight sharing, which greatly reduces the For example, the Atmel SAM D21 series microcontroller
memory requirements compared to fully connected layers. contains an ARM Cortex-M0+ CPU running at up to 48MHz,
Consequently, the FLASH footprint of the CNN and CNN- and has up to 256kB Flash and 32kB SRAM, which is
RNN models are much less. Including the recurrent layer sufcient for the model requirements given.
also requires only a minimal amount of additional memory,
namely the additional full connection from the CNN output V. DISCUSSION
to the recurrent layer input, and the amount of neurons in While deep learning approaches provide state-of-the-art
the recurrent layer. This increase in memory requirement is results in many elds, our models do not perform better
a small trade-off for the large increase in overall accuracy. than the results in [9]. There may be multiple reasons for
the lack of performance. First, we wish to maximize model
TABLE II accuracy while ensuring low memory requirements. Using
C ONFUSION M ATRIX FOR THE HAART DATASET USING CNN additional layers, increasing the number of output channels,
ARCHITECTURE . and decreasing the time dimension of each kernel in the
CNN may improve accuracy, but would increase memory
Predicted Class requirements, making the model unsuitable for in-material
True Class A B C D E F G processing. Second, RNN layers suffer from an inability
(A) No Touch 92.9 0.0 0.6 1.7 0.0 2.6 2.1 to capture long-term dependencies in a sequence; using
(B) Constant 2.9 85.8 5.4 4.8 0.1 1.0 0.1
(C) Pat 6.1 3.1 43.9 13.3 1.7 25.3 6.7 long short-term memory (LSTM, [17]) layers could improve
(D) Rub 0.3 6.3 14.2 42.9 16.5 13.8 6.1 results, though at the expense of more parameters. A limiting
(E) Scratch 0.0 0.6 6.1 7.2 60.1 8.3 17.7 factor to implementing these, however, is the possibility of
(F) Stroke 13.0 0.9 9.9 26.4 5.0 37.3 7.4
(G) Tickle 0.5 0.8 7.5 3.4 52.8 4.3 30.7 overtting due to a large number of model parameters. Aug-
menting the data by collecting additional gesture samples,
2319
TABLE V
C ONFUSION M ATRIX FOR THE C O ST DATASET USING CNN ARCHITECTURE .
Grab Hit Massage Pat Pinch Poke Press Rub Scratch Slap Squeeze Stroke Tap Tickle
Grab 60.2 2.8 8.2 0.6 4.1 0.6 2.5 1.0 1.1 0.1 14.0 1.7 0.3 2.7
Hit 3.8 22.8 1.0 10.3 14.8 9.5 2.8 1.5 0.5 7.0 4.8 5.8 5.0 10.5
Massage 5.1 0.7 46.3 0.5 6.7 0.2 2.4 13.4 2.1 0.5 6.1 4.4 0.1 11.6
Pat 2.4 6.2 2.1 45.9 1.8 1.3 7.3 2.8 1.3 6.0 1.6 3.4 10.2 7.8
Pinch 3.1 4.0 5.8 1.9 60.0 2.7 2.4 1.3 0.2 0.5 9.8 3.2 0.3 4.9
Actual Class
Poke 1.9 6.9 2.5 3.2 21.9 31.6 3.8 1.3 1.2 0.7 1.2 2.7 11.4 9.6
Press 6.4 2.7 3.3 2.7 5.8 6.9 50.0 5.0 0.3 1.2 10.8 2.0 0.7 2.2
Rub 1.1 1.5 11.9 1.4 2.1 0.5 4.2 39.2 4.2 0.6 1.5 14.0 0.4 17.5
Scratch 0.7 1.5 6.2 2.5 1.5 0.9 0.7 14.6 17.9 0.7 1.2 5.3 1.3 45.0
Slap 3.4 22.0 0.6 12.4 5.3 1.2 5.6 3.1 1.5 19.2 5.3 4.3 2.2 14.0
Squeeze 38.5 1.7 15.4 0.6 15.4 0.4 1.0 0.8 0.1 0.2 21.2 2.0 0.1 2.6
Stroke 1.9 2.7 5.4 1.3 4.5 0.1 2.1 28.8 4.0 0.7 2.1 25.1 0.5 21.0
Tap 1.5 6.9 1.8 21.5 7.1 2.7 6.0 1.5 1.6 3.8 3.6 3.5 21.2 17.2
Tickle 0.5 1.0 5.4 2.0 3.7 1.0 0.2 4.5 5.5 0.4 0.7 3.7 1.0 70.5
TABLE VI
C ONFUSION M ATRIX FOR THE C O ST DATASET USING CNN-RNN ARCHITECTURE .
Grab Hit Massage Pat Pinch Poke Press Rub Scratch Slap Squeeze Stroke Tap Tickle
Grab 73.0 0.0 1.6 0.8 0.8 0.0 0.0 0.0 3.3 0.0 17.2 3.3 0.0 0.0
Hit 0.0 1.7 0.0 94.2 1.7 0.0 0.0 0.0 0.0 0.8 0.8 0.8 0.0 0.0
Massage 0.0 0.0 67.6 2.8 1.1 0.0 0.0 2.8 7.8 0.0 2.2 11.7 0.0 3.9
Pat 0.0 0.0 0.0 91.9 0.0 0.0 0.0 0.0 0.0 2.4 0.0 2.4 0.0 0.8
Pinch 1.7 0.0 3.3 4.2 70.0 0.8 0.8 0.0 0.0 0.0 18.3 0.8 0.0 0.0
Actual Class
Poke 0.0 0.0 0.0 45.0 10.0 34.2 0.0 0.0 0.0 0.0 4.2 0.0 5.0 1.7
Press 14.2 0.0 0.0 0.8 4.2 4.2 63.3 0.0 0.0 0.0 10.8 2.5 0.0 0.0
Rub 0.0 0.0 9.2 0.0 0.0 0.0 3.8 22.3 13.8 0.0 2.3 42.3 0.0 6.2
Scratch 0.0 0.0 10.9 5.1 1.4 0.0 0.0 10.1 31.9 0.0 1.4 16.7 0.0 22.5
Slap 0.0 0.0 0.0 85.8 0.8 0.0 0.0 0.0 0.8 10.0 0.8 1.7 0.0 0.0
Squeeze 50.4 0.0 0.8 0.0 10.1 0.0 0.0 0.0 0.0 0.0 37.8 0.8 0.0 0.0
Stroke 0.0 0.0 0.8 0.8 0.8 0.0 0.0 2.5 0.0 0.8 0.0 88.4 0.0 5.8
Tap 0.0 0.0 0.0 82.5 0.0 0.8 0.8 0.0 0.8 0.0 0.0 0.0 15.0 0.0
Tickle 0.0 0.0 8.6 4.6 1.3 0.0 0.0 1.3 0.7 0.0 0.7 1.3 2.0 75.5
TABLE VII
C ONFUSION M ATRIX FOR THE C O ST DATASET USING AUTOENCODER -RNN ARCHITECTURE .
Grab Hit Massage Pat Pinch Poke Press Rub Scratch Slap Squeeze Stroke Tap Tickle
Grab 29.2 0.0 18.2 0.0 0.0 0.7 0.0 0.0 0.0 0.0 47.4 0.7 0.0 3.6
Hit 0.0 7.5 0.8 0.0 4.2 21.6 0.0 0.0 0.0 2.5 0.0 0.8 39.2 23.3
Massage 2.3 1.3 57.4 0.0 3.5 1.3 0.0 11.9 1.6 0.3 3.5 6.1 0.6 10.0
Pat 0.0 5.8 0.8 5.8 3.3 2.5 0.0 0.8 0.0 9.2 0.0 5.8 25.8 40.0
Pinch 0.0 0.8 26.8 0.0 21.1 17.9 0.0 0.0 2.4 0.8 4.9 4.1 0.8 20.3
Actual Class
Poke 0.0 7.4 0.0 0.0 9.9 59.5 0.0 0.0 0.0 0.0 0.0 0.0 4.1 19.0
Press 9.0 5.7 9.8 1.6 1.6 4.1 39.3 9.8 4.1 1.6 5.7 4.1 1.6 1.6
Rub 2.1 1.5 18.6 0.5 0.5 1.5 13.9 22.2 2.6 4.1 0.0 10.3 3.6 18.6
Scratch 0.5 2.4 9.7 3.4 2.4 3.4 1.9 5.8 13.1 1.0 0.5 10.2 4.9 40.8
Slap 0.0 5.8 0.0 0.0 3.3 10.0 0.0 0.0 0.0 5.8 0.0 0.0 59.2 15.8
Squeeze 19.7 0.7 23.5 0.0 3.0 1.5 0.0 0.0 0.0 0.0 47.7 0.0 2.3 2.3
Stroke 0.0 0.0 6.9 0.0 4.9 1.4 0.0 11.1 2.1 2.1 0.0 37.5 2.8 31.3
Tap 0.0 4.9 0.8 4.1 5.8 8.3 0.0 0.0 0.0 5.8 0.0 0.8 29.8 39.7
Tickle 0.0 0.4 8.4 0.0 5.9 5.5 0.0 0.0 2.5 0.0 0.0 5.9 3.8 67.6
adding noise, or rotating or translating samples may mitigate robot, rather than new users. In these applications, we would
this issue and provide improved results. expect to see an increase in model performance, as models
can learn gestures specic to one or a few subjects, rather
Splitting the data based on subject for training and test than needing to generalize to a large population using gesture
sets introduces subject dependence in the classier. Model information from a few individuals.
performance is expected to be reduced in this scenario, when
compared to splitting the data based on individual gestures. VI. CONCLUSION
In many applications, such as with assistive or companion We present three deep learning models for the task of
robots, we expect the same user to be interacting with the identifying affective touch for social robots: a CNN model,
2320
TABLE VIII
[4] M. M. Jung, R. Poppe, M. Poel, and D. K. J. Heylen, Touching
M EMORY REQUIREMENTS FOR EACH MODEL . the voidintroducing cost: Corpus of social touch, in Proceedings of
the 16th International Conference on Multimodal Interaction. ACM,
2014, pp. 120127.
FLASH RAM
[5] M. M. Jung, X. L. Cang, M. Poel, and K. E. MacLean, Touch
CNN 15,714 6,414 (44,994)
HAART CoST
challenge15: Recognizing social touch gestures, in Proceedings of
CNN-RNN 18,264 6,464 (45,044)
the 2015 ACM on International Conference on Multimodal Interaction.
Autoencoder-RNN 39,949 488
ACM, 2015, pp. 387390.
CNN 17,777 4,255 (24,095)
[6] X. L. Cang, P. Bucci, A. Strang, J. Allen, K. MacLean, and H. Liu,
CNN-RNN 20,327 4,305 (24,145)
Different strokes and different folks: Economical dynamic surface
Autoencoder-RNN 39,767 481
sensing and affect-related touch recognition, in Proceedings of the
2015 ACM on International Conference on Multimodal Interaction.
ACM, 2015, pp. 147154.
[7] T. Balli Altuglu and K. Altun, Recognizing touch gestures for
a CNN-RNN model and an Autoencoder-RNN model. These social human-robot interaction, in Proceedings of the 2015 ACM on
International Conference on Multimodal Interaction. ACM, 2015,
model provide similar performance in terms of accuracy pp. 407413.
as those presented in the 2015 Social Touch Challenge. [8] Y. F. A. Gaus, T. Olugbade, A. Jan, R. Qin, J. Liu, F. Zhang, H. Meng,
While none of the models outperform the best models in and N. Bianchi-Berthouze, Social touch gesture recognition using
random forest and boosting on distinct feature sets, in Proceedings of
the Social Touch Challenge in terms of overall accuracy, the the 2015 ACM on International Conference on Multimodal Interaction.
deep learning models explored here provide a touch gesture ACM, 2015, pp. 399406.
prediction in real-time at a rate of 6 to 9 times per second, [9] V.-C. Ta, W. Johal, M. Portaz, E. Castelli, and D. Vaufreydaz,
The grenoble system for the social touch challenge at icmi 2015,
while the models in the Social Touch Challenge only provide in Proceedings of the 2015 ACM on International Conference on
predictions once the entire gesture is captured. Additionally, Multimodal Interaction, 2015, pp. 391398.
our models automatically extract meaningful features from [10] D. Hughes, N. Farrow, H. Prota, and N. Correll, Detecting and iden-
tifying tactile gestures using deep autoencoders, geometric moments
training data, rather than requiring hand-engineered features and gesture level features, in Proceedings of the 2015 on International
to be designed and evaluated. Conference on Multimodal Interaction. ACM, 2015, pp. 415422.
The models we consider can provide high prediction rates [11] M. A. McEvoy and N. Correll, Materials that couple sensing,
actuation, computation and communication, Science, vol. 347, no.
based on the hop size of the sampled windows. The datasets 6228, p. 1261689, March 2015.
used were collected for social touch classication, where [12] D. Hughes and N. Correll, Distributed machine learning in materials
applications involved determining the emotion conveyed by that couple sensing, actuation, computation and communication,
arXiv preprint arXiv: 1606.03508, June 2016.
a human subject and rapid response is not critical. In other [13] W. D. Stiehl, J. Liberman, C. Breazeal, L. Basel, L. Lalla, and M. Wolf,
applications, quick response to a touch gesture may be Design of a therapeutic robotic companion for relational, affective
important and justify a lower classication accuracy, such as touch, in Proceedings of the International Workshop on Robot and
Human Interactive Communication (ROMAN), 2005, pp. 408415.
with assistive robots or providing tactile-based commands in [14] S. Ji, W. Xu, M. YAng, and K. Yu, 3d convolutional neural net-
industrial settings. works for human activity recognition, IEEE Transactions on Pattern
The memory requirements for these models were also eval- Analysis and Machine Intelligence, vol. 35, no. 1, pp. 221231, 2013.
[15] G. E. Hinton and R. R. Salakhutdinov, Reducing the dimensionality
uated to determine if deep learning approaches are suitable of data with neural networks, Science, vol. 313, pp. 504507, 2006.
for gesture recognition in a Robotic Materials context. We [16] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro,
have determined that the models could be implemented on G. S. Corrado, A. Davis, J. Dean, M. Devin, S. Ghemawat,
I. Goodfellow, A. Harp, G. Irving, M. Isard, Y. Jia, R. Jozefowicz,
inexpensive microcontrollers, though the memory consump- L. Kaiser, M. Kudlur, J. Levenberg, D. Mane, R. Monga,
tion would be a signicant portion of the available memory, S. Moore, D. Murray, C. Olah, M. Schuster, J. Shlens, B. Steiner,
demonstrating that deep learning approaches are promising I. Sutskever, K. Talwar, P. Tucker, V. Vanhoucke, V. Vasudevan,
F. Viegas, O. Vinyals, P. Warden, M. Wattenberg, M. Wicke,
for Robotic Materials. The memory requirements for the Y. Yu, and X. Zheng, TensorFlow: Large-scale machine learning on
models presented in the Social Touch Challenge are not heterogeneous systems, 2015, software available from tensorow.org.
known, so we are unable to compare the requirements of our [Online]. Available: http://tensorow.org/
[17] S. Hochreiter and J. Schmidhuber, Long short-term memory, Neural
models with those. However, as the entire gesture is needed Computation, vol. 9, no. 8, pp. 17351780, 1997.
to extract the features, we can estimate the RAM needed
to store an entire gesture to be 55kB, implying that the
approaches used in the Social Touch Challenge may not be
suitable for a Robotic Material approach.
R EFERENCES
[1] R. S. Dahiya, P. Mittendorfer, M. Valle, G. Cheng, and V. J. Lumelsky,
Directions toward effective utilization of tactile skin: A review, IEEE
Sensors Journal, vol. 13, no. 11, pp. 41214138, 2013.
[2] S. Yohanan and K. E. MacLean, The role of affective touch in human-
robot interaction: Human intent and expectations in touching the haptic
creature, International Journal of Social Robotics, vol. 4, no. 2, pp.
163180, 2012.
[3] A. Flagg and K. MacLean, Affective touch gesture recognition for a
furry zoomorphic machine, in Proceedings of the 7th International
Conference on Tangible, Embedded and Embodied Interaction. ACM,
2013, pp. 2532.
2321