Robot Auditory Interface Design Factor Attributes and the Level Values Based on UX using TTS System

  • Authors

    • Seung Eun Chung
    • Han Young Ryoo
    https://doi.org/10.14419/ijet.v8i1.4.25457
  • Social Service Robot, Auditory Interface, Design Factors and values, User Experience
  • The purpose of this study is to suggest the attributes of the design factors and the level values for each attribute for the service robot’s auditory interface design. To do so, the main design factors of the auditory interface were organized through the research on the existing literature, and then the attributes and the level values of the auditory interface design factors that can be designed using the SSML from the TTS system were organized. A survey was conducted to verify whether the organized values have a significance meaning in the aspect of the user experience, and the level values which had differences in all the functional/service experience, the interaction experience, and the emotional experience have been organized. From the results, ‘Prototype of the sound effect’, ‘Gender’, ‘Pitch’, ‘Pitch range’, and ‘Rate’ were verified to be useable as the attributes of the auditory interface design factors to measure the overall user experience with the robots. Two out of the 15 investigated subordinate level values had been deleted and integrated so that a total of 13 level values have been suggested in the end. This research has a significant meaning in a way that it offers suggestions to the designers of the robots that provide auditory interface through the TTS system the usable design factors for the user experience-centered designs.

     

  • References

    1. [1] Yoshimura T, Tokuda K, Masuko T, Kobayashi T & Kitamura T (1999), Simultaneous Modeling of Spectrum, Pitch and Duration in HMM-Based Speech Synthesis, Proceeding of EUROSPEECH 5, 2347-2350.

      [2] Lee HJ (2015). Korean Sentence Symbol Preprocess System for the Improvement of Speech Synthesis Quality. Journal of the Korea Society of Computer and Information 20(2), 149-156.

      [3] Kim Y, Ahn S & Lee T (2017) UX Guidelines for Designing Audio Guidance of Multimedia Contents for Low-Vision - With a Focus on Sound Level Difference and Amount of Auditory Information, Archives of Design Research 30(1), 131-142.

      [4] Asakawa C, Takagi H, Ino S & Ifukube T (2003). Maximum listening speeds for the blind. Proceedings of the 9th International Conference on Auditory Display (ICAD).

      [5] Kwak S, Goh T, Park K & Ahn J (2012). The Usability of a Robot and Human Empathy by Anthropomorphic Sound Feedback. Journal of Korean Society of Design Science 23(3), 20-27.

      [6] Yang B (1997). A Study on the Human Auditory Scaling. Voice science 2, 125-134.

      [7] Park HW, Jee SH, & Bae MJ (2016). Study on the Confidence-Parameter Estimation through Speech Signal. Asia-pacific Journal of Multimedia Services Convergent witth Art, Humanities, and Sociology 6(7), 101-108.

      [8] Mehrabian A (1971). Silent Messages. Belmont, Ca: Wadsworth Publishing Company.

      [9] Arons B & Mynatt E (1994). The future of speech and audio in the interface: a CHI'94 workshop. Proceedings of ACM SIGCHI Bulletin, CHI'94 Reports 26(4), 44-48.

      [10] Garzonis S, Jones S, Jay T, & O'Neill E (2009). Auditory icon and earcon mobile service notifications: intuitiveness, learnability, memorability and preference. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM. 1513-1522

      [11] Stephen AB, Peter CW, & Alistair DNE (1993). An evaluation of earcons for use in auditory human-computer interfaces. Proceedings of the INTERACT'93 and CHI'93 conference on Human factors in computing systems. ACM. 222-227.

      [12] Vilimek R & Hempel T (2005). Effects of speech and non-speech sounds on short-term memory and possible implications for in-vehicle use. Proceedings of ICAD 05-Eleventh Meeting of the International Conference on Auditory Display. International Community for Auditory Display. 5. 344-350

      [13] Chae HS, Hong JY, Jeon MH, & Han KH (2007), A Study on Voice User Interface for Domestic Appliance. Korean Journal of the Science of Emotion and Sensibility 10(1), 55-68.

      [14] Jee ES, Jeong YJ, Kim CH, & Kobayashi H (2010). Sound design for emotion and intention expression of socially interactive robots. Intelligent Service Robotics 3(3), 199-206.

      [15] Choi JH, Cho DU, & Jeong YM (2016). Identification of voice for listeners who feel favor using voice analysis. The Journal of Korean Institute of Communications and Information Sciences 41(1), 122-131.

      [16] Schmitz M, Krüger A, & Schmidt S (2007). Modelling personality in voices of talking products through prosodic parameters. Proceedings of the 12th international conference on Intelligent user interfaces. ACM. 313-316.

      [17] Cho DU (2015), Voice feature analysis of next president candidate in south korea. Proceedings of KICS Summer Conference 509-510.

      [18] Laukka P, Juslin PN, & Bresin R (2005), A dimensional approach to vocal expression of emotion. Cognition and Emotion 19(5), 633–653

      [19] Cho DU (2015), Voice feature analysis of CEO of the 3 major companies in south korea by applying IT Technologies, Proceedings of KICS Summer Conference

      [20] Apple W, Streeter LA, & Krauss RM (1979), Effects of pitch and speech rate on personal attributions. Journal of Personality and Social Psychology 37, 715–727.

      [21] Choi Y, Jung Y, Kim Y, Suh Y, & Kim H (2018), An end-to-end synthesis method for Korean text-to-speech systems, Phonetics and Speech Sciences 10(1), 39-48.

      [22] Lee JS (2003), Design and Implementation of a Speech Synthesis Markup Language Processing System for Korean, Dissertation ,Graduate School of Sungshin Women's University.

      [23] W3C (2010), Speech Synthesis Markup Language (SSML) Version 1.1, available online: https://www.w3.org/TR/2010/REC-speech-synthesis11-20100907/ last visit: 01.02. 2018

      [24] Chung SE & Ryoo HY (2018), Aural and Behavioral Factor Research for Robot Design, Proceedings of International Journal of Multimedia and Ubiquitous Engineering 13(2), 1-6.

  • Downloads

  • How to Cite

    Eun Chung, S., & Young Ryoo, H. (2019). Robot Auditory Interface Design Factor Attributes and the Level Values Based on UX using TTS System. International Journal of Engineering & Technology, 8(1.4), 471-477. https://doi.org/10.14419/ijet.v8i1.4.25457