Article published In:
Human Robot Collaborative Intelligence: Theory and applications
Edited by Chenguang Yang, Xiaofeng Liu, Junpei Zhong and Angelo Cangelosi
[Interaction Studies 20:1] 2019
► pp. 2541
References (25)
References
Brady, K., Y. Gwon, P. Khorrami, E. Godoy, W. Campbell, C. Dagli, and T. S. Huang. (2016). “Multi-modal audio, video and physiological sensor learning for continuous emotion prediction,” in International Workshop on Audio/visual Emotion Challenge, pp. 97–104. DOI logoGoogle Scholar
Chang, C. C. and C. J. Lin. (2011). “Libsvm: A library for support vector machines,” vol. 2, no. 3, pp. 1–27. DOI logoGoogle Scholar
Chen, Z. and B. Liu. (2014). “Topic modeling using topics from many domains, lifelong learning and big data,” in International Conference on International Conference on Machine Learning, pp. II–703.Google Scholar
. (2016). “Lifelong machine learning,” Synthesis Lectures on Artificial Intelligence & Machine Learning, vol. 10, no. 3, pp. 1–145. DOI logoGoogle Scholar
Chen, Z., N. Ma, and B. Liu. (2018). “Lifelong learning for sentiment classification,” arXiv preprint arXiv, vol. 21, no. 7, pp. 845–853.Google Scholar
Debrot, A., D. Schoebi, M. Perrez, and A. B. Horn. (2013). “Touch as an interpersonal emotion regulation process in couples’ daily lives: the mediating role of psychological intimacy,” Personality & Social Psychology Bulletin, vol. 39, no. 10, p. 1373. DOI logoGoogle Scholar
Gao, Y., N. Bianchi-Berthouze, and H. Meng. (2012). “What does touch tell us about emotions in touchscreen-based gameplay?ACM Transactions on Computer-Human Interaction (TOCHI), vol. 19, no. 4, pp. 1–30. DOI logoGoogle Scholar
Graesser, A.. (2011). “The half-life of cognitive-affective states during complex learning,” Cognition & Emotion, vol. 25, no. 7, p. 1299. DOI logoGoogle Scholar
Guest, S., J. M. Dessirier, A. Mehrabyan, F. Mcglone, G. Essick, G. Gescheider, A. Fontana, R. Xiong, R. Ackerley, and K. Blot. (2011). “The development and validation of sensory and emotional scales of touch perception.” Attention Perception & Psychophysics, vol. 73, no. 2, pp. 531–550. DOI logoGoogle Scholar
Hertenstein, M. J., R. Holmes, M. Mccullough, and D. Keltner. (2009). “The communication of emotion via touch.” Emotion, vol. 9, no. 4, p. 566. DOI logoGoogle Scholar
Huang, G. B., D. H. Wang, and Y. Lan. (2011). “Extreme learning machines: a survey,” International Journal of Machine Learning & Cybernetics, vol. 2, no. 2, pp. 107–122. DOI logoGoogle Scholar
Jung, M. M., R. Poppe, M. Poel, and D. K. J. Heylen. (2014). “Touching the void – introducing cost:corpus of social touch,” in Icmi 14 International Conference on Multimodal Interaction, pp. 120–127.Google Scholar
Jung, M. M., M. Poel, R. Poppe, and D. K. J. Heylen. (2017). “Automatic recognition of touch gestures in the corpus of social touch,” Journal on Multimodal User Interfaces, vol. 11, no. 1, pp. 81–96. DOI logoGoogle Scholar
Liu, H., J. Qin, F. Sun, and D. Guo. (2017). “Extreme kernel sparse learning for tactile object recognition,” IEEE Transactions on Cybernetics, vol. 47, no.12, pp., 4509–4520. DOI logoGoogle Scholar
Liu, H., F. Sun, D. Guo, and B. Fang. (2017). “Structured output-associated dictionary learning for haptic understanding,” IEEE Transactions on Systems, Man and Cybernetics: Systems, vol.47, no.7, pp.1564–1574. DOI logoGoogle Scholar
Maramis, C., L. Stefanopoulos, I. Chouvarda, and N. Maglaveras. (2018). “Emotion Recognition from Haptic Touch on Android Device Screens.” Precision Medicine Powered by pHealth and Connected Health, pp. 205–209. DOI logoGoogle Scholar
Matsuda, Y., T. Isomura, I. Sakuma, Y. Jimbo, E. Kobayashi, and T. Arafune. (2008). “Emotion recognition of finger braille,” in International Conference on Intelligent Information Hiding and Multimedia Signal Processing, pp. 1408–1411. DOI logoGoogle Scholar
Morrison, I., L. S. Loken, and H. Olausson. (2010). “The skin as a social organ.” Experimental Brain Research, vol. 204, no. 3, pp. 305–314. DOI logoGoogle Scholar
Park, Y. W., K. M. Baek, and T. J. Nam. (2013). “The roles of touch during phone conversations:long-distance couples’ use of poke in their homes,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 1679–1688. DOI logoGoogle Scholar
Preeti, K. and M. Sasikumar. (2010). “Recognising emotions from keyboard stroke pattern,” International Journal of Computer Applications, vol. 11, no. 9, pp. 24–28.Google Scholar
Ruvolo, P. and E. Eaton. (2013). “Ella: an efficient lifelong learning algorithm,” In International Conference on International Conference on Machine Learning, pp. I–507.Google Scholar
Silver, D. L., Q. Yang, and L. Li. (2013). “Lifelong machine learning systems: Beyond learning algorithms,” in AAAI 2013 Spring Symposium on Lifelong Machine Learning.Google Scholar
Thrun, S. and T. M. Mitchell. (1995). “Lifelong robot learning,” Robotics & Autonomous Systems, vol. 15, no. 1–2, pp. 25–46. DOI logoGoogle Scholar
Yang, C., K. Huang, H. Cheng. (2017). “Haptic identification by ELM-controlled uncertain manipulator,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, 47(8): 2398–2409. DOI logoGoogle Scholar
Yang, C., C. Zeng, P. Liang, Z. Li, R. Li, C. Su. (2018). “Interface Design of a Physical Human-Robot Interaction System for Human Impedance Adaptive Skill Transfer,” IEEE Trans. Automation Science and Engineering 15(1): 329–340. DOI logoGoogle Scholar
Cited by (3)

Cited by three other publications

Liu, Man
2022. English speech emotion recognition method based on speech recognition. International Journal of Speech Technology 25:2  pp. 391 ff. DOI logo
Li, Zhen, Jiao Zhang, Mengwan Li, Jizhuo Huang & Xiangyu Wang
2020. A Review of Smart Design Based on Interactive Experience in Building Systems. Sustainability 12:17  pp. 6760 ff. DOI logo
Liu, Huaping, Di Guo, Fuchun Sun, Wuqiang Yang, Steve Furber & Tengchen Sun
2020. Embodied tactile perception and learning. Brain Science Advances 6:2  pp. 132 ff. DOI logo

This list is based on CrossRef data as of 18 october 2024. Please note that it may not be complete. Sources presented here have been supplied by the respective publishers. Any errors therein should be reported to them.