<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">First Workshop on Social Interaction and Multimodal Expression for Socially Intelligent Robots</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Christiana</forename><surname>Tsiourti</surname></persName>
							<email>christiana.tsiourti@unige.ch</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Service Science</orgName>
								<orgName type="institution">University of Geneva</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jorge</forename><surname>Dias</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Institute of Systems</orgName>
								<orgName type="institution">Robotics</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">University of Coimbra</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Astrid</forename><surname>Weiss</surname></persName>
							<email>astrid.weiss@tuwien.ac.at</email>
							<affiliation key="aff3">
								<orgName type="department">CIN Institute of Automation</orgName>
								<orgName type="institution">Control Vienna University of Technology</orgName>
								<address>
									<settlement>Vienna</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sten</forename><surname>Hanke</surname></persName>
							<email>sten.hanke@ait.ac.at</email>
							<affiliation key="aff4">
								<orgName type="institution">AIT Austrian Institute of Technology GmbH Center of Health and Bioresources</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Julian</forename><surname>Angel-Fernandez</surname></persName>
							<affiliation key="aff3">
								<orgName type="department">CIN Institute of Automation</orgName>
								<orgName type="institution">Control Vienna University of Technology</orgName>
								<address>
									<settlement>Vienna</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Luis</forename><surname>Santos</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Institute of Systems</orgName>
								<orgName type="institution">Robotics</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">University of Coimbra</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">First Workshop on Social Interaction and Multimodal Expression for Socially Intelligent Robots</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">4818543FBD77EB61FE20AA94DE5FD117</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T22:50+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Recent advances in the field of robotics contributed to the development of several kinds of robots able to express social communication skills using a number of interactive modalities such as facial expressions, gestures, gaze, motion, and color. Despite this progress, the multimodal expression capabilities of robots are still far behind the intuitiveness and naturalness that is required to allow uninformed people to interact in their everyday life with naturally communicative robots. The workshop on "Social Interaction and Multimodal Expression for Socially Intelligent Robots" is held in conjunction with the IEEE International Symposium on Robot and Human Interactive Communication. Its topics cover interdisciplinary research on understanding, designing, and evaluating multimodal communication skills for socially intelligent robots that have the potential to enhance our daily lives in our homes and at work. This years workshop consists of four talks covering topics such as verbal and non-verbal interaction with robots, design of affective expressions and applications for multimodal interaction with social robots. Moreover, it includes three keynote talks by Ana Paiva, Andrea Bonarini and Jorge Dias and a brainstorming discussion.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Workshop Overview</head><p>In the last few years, a growing interest has been seen in the development of autonomous interactive robots to enhance our daily lives in our homes and at work. Applications are plentiful: companions for children or elderly people, partners in industries, guides in public or personal spaces, educational tutors at school and so on. Robots able to face such broad range of social situations require social-cognitive capabilities that promote fluid and effective interactions. These include the automatic understanding of the users actions, behaviors, and mental and emotional states and the coherent production of multimodal, verbal and non-verbal communication skills.</p><p>Designing and developing multimodal communication skills for robots in order to provide more natural and powerful interactive experiences is a significant challenge in practice due to limitations both in technology and in our understanding of how different modalities must work together to convey human-like levels of social intelligence. The area of social interaction and multimodal expression for socially intelligent robots remains very much an active research area with many challenges and open research questions. For example, to what degree, and how precisely, different modalities (e.g., eye gaze, touch, vocal, body, and facial expressions) might be involved in human interaction with intelligent robots remains largely unknown. Likewise, whether combining multimodal elements of emotional expressions result in enhanced recognition of the emotion is still an open question. Furthermore, current research is predominantly focused on visual and auditory senses, neglecting other modalities such as touch that are an integral part of how we experience the real world around us.</p><p>The scope of this workshop is to present rigorous scientific advances on social interaction and multimodal expression for socially intelligent robots. Previous research findings show that this challenge cannot be solely approached from a pure engineering perspective. Human sciences, social sciences, and cognitive sciences play a primary role in the development and the enhancement of social interaction skills for socially intelligent robots. This workshop fosters interdisciplinary collaboration between researchers on the domain, addressing both the study of human-human interactions as well as human-machine interactions. The analysis of human-human interactions offers researchers the opportunity to understand how humans interact with the world and with one another multimodally, through both parallel and sequential use of multiple modalities (e.g., eye gaze, touch, vocal, body, and facial expressions), and to develop guidelines on how to design robot behaviors. On the other hand, results achieved by researchers studying human-robot interactions are particularly important to understand how the multimodal communication skills developed for social robots are perceived by uninformed interaction partners (e.g., children, elderly) and how they influence the interaction process (e.g., regarding usability and acceptance).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Submission Categories</head><p>In order to enable a diverse program, the workshop invites a variety of submission categories in the call for papers. Topics included, but where not limited to:</p><p>(a) Contributions of fundamental nature -Psychophysical studies and empirical research about multimodality (b) Technical contributions on multimodal interaction -Novel strategies of multimodal human-robot interactions -Dialogue management using multimodal output -Work focusing on novel modalities (e.g., touch) (c) Multimodal interaction evaluation -Evaluation and benchmarking of multimodal human-robot interactions -Empirical HRI studies with (partial) functional systems -Methodologies for the recording, annotation, and analysis of multimodal interactions (d) Applications for multimodal interaction with social robots -Novel application domains for multimodal interaction (e) Position papers and reviews of the state-of-the-art and ongoing research</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Program Overview</head><p>The workshop consists of three invited talks and four talks based on paper submissions covering topics such as verbal and non-verbal interaction with robots, design of affective expressions and applications for multimodal interaction with social robots. The talks were followed by a brainstorming discussion. Table <ref type="table" target="#tab_0">1</ref> presents an outline of the program.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Invited Speakers</head><p>Ana Paiva -Instituto Superior Tcnico, Technical University of Lisbon Short Bio: Ana Paiva is an Associate Professor in the Department of Computer Science and Engineering (Departamento de Engenharia Informtica) of Instituto Superior Tcnico from the Technical University of Lisbon (Universidade Tcnica de Lisboa). She is also the group leader of GAIPS (Grupo de Agentes Inteligentes e Personagens Sintticas), a research group on agents and synthetic characters at INESC-ID. Her main scientific interests lay in the area of Autonomous Agents, Embodied Conversational Agents and Robots and Multiagent Simulation Systems. Prof. Ana Paiva has been researching in the area of artificial intelligence for the past twenty years, having also taught at IST during that period.</p><p>Talk: Robots that listen to the users heart: The role of emotions in multimodal communication with social robots Andrea Bonarini -Department of Electronics, Information, and Bioengineering, Politecnico di Milano Short Bio: Andrea Bonarini is full professor and coordinator of the AI and Robotics Lab at Politecnico di Milano. His current research interests are on machine learning and emotional human-robot interaction, with special focus on robotic games and toys for entertainment and social, psychological and physical development of people with and without special needs. Since 1989, he has realized with his collaborators and students more than 60 autonomous robots. In 2015, he co-funded Nova Labs, to share experience and tools to produce professional robots in a short time at low cost.</p><p>Talk: Speechless Social Robotic Toys: Emotional Relations Without Words Abstract: More and more sophisticated toys that can perceive signals from the world and physically act in consequence, i.e. robotic toys, are entering the market and are initiating a new way of playing: they cannot any longer treated as props (as traditional plushes and dolls are), but they are agents that can decide their actions, to interact with. This arrangement opens challenges concerning believability and appropriateness of the robots behavior, which cannot really exploit speech because of both technological and privacy reasons. Multimodal interaction is relevant for this application, and should keep into account many dimensions, from psychology, to technology, and cost. Talk: BUM -Bayesian User Model for Distributed Social Robots Abstract: In this work we present a Bayesian User Model for inferring the characteristics and inter-personal relationships of a population users. The model can receive evidence gathered by various interactive devices, such as social robots or wearable devices. The system is modular, with each module being responsible for gathering information and observations from persons present in the systems operation scenario. This information enables each module to determine a single characteristic of the person. New observations and measurements received by the system are fused with previous knowledge by a sub-process based on an information theory technique. This allows the system to be implemented in diverse heterogeneous distributed system topologies, extending beyond robotics. We have conducted experiments involving a simulated team of social robots and user population with four sets of person types. Our experiments have shown that the system is able to learn and classify the persons characteristics, and to find relevant user groups via clustering. This system can potentially be used to gather information on a large set of persons, as well as to be an information source for user-adaptive applications in areas such as Robotics, AAL and Internet of Things. Astrid Weiss is a postdoctoral research fellow in HRI at the Vision4Robotics group at the ACIN Institute of Automation and Control at Vienna University of Technology (Austria). She holds a masters degree in sociology and a PhD in social sciences from the University of Salzburg. Her current research focuses on user-centered design and evaluation studies for Human-Robot Interaction, with a special interest in the impact technology has on the everyday life and what makes people accept or reject technology. She co-organized workshops on a variety of HRI-related topics at the following conferences: RO-MAN2008, HRI2009, HRI2011, ICSR2013, and HRI2014. Moreover, she is regularly member of Program and Organizing Committees related to HRI research.</p><p>Sten Hanke has a Master in electrical engineering and a PhD in medical science. He is a researcher for the Austrian Institue of Technology (AIT) biomedical engineering in smart biomedical systems. He joined AIT in the year 2005. He is working on European and national projects in the field of ambient assisted living, social robotics, interaction with ICT systems, smart systems as well as psycho-physiological assessments since more than 10 years. At the moment is he technical manager of the FP7 Miraculous-Life project<ref type="foot" target="#foot_5">10</ref> dealing with development of a virtual support partner as well as workpackage leader of the H2020 GrowMeUp project<ref type="foot" target="#foot_6">11</ref> developing an ICT service robot for ambient assisted living environments. He is a working member of the international ISO/IEEE 11073, European ISO/TC 215 and CEN/TC 251 health care and medical device communication working group as well as the national ON-AG238.4 group. Currently he is in the Ambient Assisted Living Open Association (AALOA) Governing Board.</p><p>Julian Angel-Fernandez is a postdoctoral research at the Visrion4Robotics at the Institute of Automation and Control (ACIN) at University of Technology (Austria). His interests twofold in representation and selection of emotions in robotics, and study of methodologies to increase the interest of students in STEM careers. Currently, he is working as scientific coordinator in the European project ER4STEM, which aims to maintain childrens curiosity in technology through the use of robotics. He holds a European PhD with merit in Information and Technology (2016), from Politecnico di Milano Italy. He also holds a bachelor in Electronic <ref type="bibr">Engineering (2008)</ref>; Bachelor in Computer Science (2009); Master in Computer Science (2010) from Universidad de los Andes, Bogota Colombia. He has worked as temporal professor at Potificia Universidad Javeriana at the department of computer engineering (Colombia). During this period, he was part of the research group Takina and taught: programming courses (I and II), as programming languages and introduction to artificial intelligence.</p><p>Luis Santos holds a PhD from the University of Coimbra (UC), Portugal. He is currently an Invited Researcher at University of Coimbra and is acting as a Project and Technical Manager for the H2020 GrowMeUp Project (GA 643647 12 . He held a postdoctoral position at the Institute of Systems and Robotics of the University of Coimbra from October 2014 to March 2015. He acted as a seconded researcher from the UC to a company, Citard Services Ltd., within the scope of the Marie Curie IAPP Project Social Robot (GA 285870 13 , till the 30th of September, 2014. He holds a Pre-Bologna M.Sc. degree in Electrical and Computer Engineering, specialization in Robotics and Automation by the Faculty of Science and Technology from the University of Coimbra. He worked as researcher in the area of human-robot interaction and action recognition using Laban Movement Analysis, within the BACS European Project from November 2007 to February 2010. His current research topics mainly focus on machine learning methods for autonomous cognitive systems and human motion analysis towards end-user applications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusion</head><p>In summary, the first workshop on "Social Interaction and Multimodal Expression for Socially Intelligent Robots" brings together a multidisciplinary audience to discuss open questions, address difficult challenges and elaborate on novel ways to advance research in the field of mutlimodal interaction, based on theories from human-human interaction and on empirical findings validated human-robot interaction studies.</p><p>Advancing the development of socially intelligent robots opens new social and economic opportunities for the application of robots in our daily lives. We hope that the papers and the outcomes of the brainstorming discussion will contribute towards the establishment of a set of standard guidelines for the production of coherent multimodal, verbal and non-verbal communication skills for robots. </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Jorge Dias -Institute of Systems and Robotics, University of Coimbra Short Bio: Jorge Dias has been Associated Professor at the University of Coimbra with activities in the Department of Electrical Engineering and Computers and the Institute of Systems and Robotics.His research activities are in the area of Computer Vision and Robotics and has contributions on the field since 1984. Jorge Dias coordinates the research group for Artificial Perception for Intelligent Systems and Robotics of Institute of Systems and Robotics from University of Coimbra and the Laboratory of Systems and Automation of the Instituto Pedro Nunes. Since July 2011, Jorge Dias is acting as Faculty from ECE/Robotics at Khalifa University in Abu Dhabi.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>is PhD student at the Institute of Service Science of the University of Geneva (Switzerland), and a member of the Swiss Doctoral School in Affective Sciences, at the Swiss Center for Affective Science. In November 2016, she has been awarded a Doc.Mobility Fellowship by the Swiss National Science Foundation, allowing her to spend one year as a visiting researcher at the Vision4Robotics group at the ACIN Institute of Automation and Control at Vienna University of Technology (Austria). Christiana holds a Masters (2011) and Bachelor (2009) in Computer Science from the University of Cyprus (Cyprus). Between 2011 and 2013, she worked as a researcher at the Institute of Systems and Robotics of the University of Coimbra. Christianas research is focused on the development and evaluation of affective socially intelligent agents (robots and avatars) that autonomously integrate into our daily life environments and possess social skills, such as the automatic understanding of the users actions and emotional states, and the production of coherent emotional feedback. Jorge Dias has a Habilitation degree and a Ph.D. on Electrical Engineering by the University of Coimbra, with specialization in Control and Instrumentation. Jorge has been Associated Professor at the University of Coimbra (UC) 5 with activities in the Department of Electrical Engineering and Computers 6 and the Institute of Systems and Robotics (ISR) 7 .Jorge does research in the area of Computer Vision and Robotics and has contributions in the field since 1984, with several publications in international journals, books, and conferences. Jorge has been teaching several courses on Computer Vision, Robotics, Automation and Electrical Engineering and Computer Science and supervised several Ph.D. and Master students in the field of Computer Vision and Robotics. He has been principal investigator in several research international projects. Jorge coordinated the Artificial Perception Laboratory at the ISR and the Laboratory of Systems and Automation (LAS) 8 at the Instituto Pedro Nunes (IPN) 9 , and was Vice-President of IPN from June 2008 to June 2011. Jorge is currently on a leave of absence from the UC and acting as Faculty in the Electrical and Computer Engineering/Robotics Department at Khalifa University in Abu Dhabi.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Final Workshop Program -August 28 th 2017</figDesc><table><row><cell cols="3">09:00 -09:05 Organizers' introduction and overview of</cell></row><row><cell>workshop goals</cell><cell></cell><cell></cell></row><row><cell cols="3">09:05 -09:50 Invited Talk 1: Jorge Dias (University of</cell></row><row><cell cols="3">Coimbra) "BUM -Bayesian User Model for</cell></row><row><cell cols="2">Distributed Social Robots"</cell><cell></cell></row><row><cell cols="3">09:50 -10:30 Bob Schadenberg, Dirk Heylen and Vanessa</cell></row><row><cell cols="3">Evers: Affect bursts to constrain the meaning</cell></row><row><cell cols="3">of the facial expression of the humanoid robot</cell></row><row><cell>Zeno</cell><cell></cell><cell></cell></row><row><cell cols="3">Raymond Cuijpers: Turn-taking cue delays in</cell></row><row><cell cols="2">human-robot communication</cell><cell></cell></row><row><cell>10:30 -10:50 Coffee break</cell><cell></cell><cell></cell></row><row><cell cols="3">10:50 -11:35 Invited Talk 2: Andrea Bonarini (Politec-</cell></row><row><cell cols="3">nico di Milano) Speechless social robotic toys:</cell></row><row><cell cols="2">emotional relations without words</cell><cell></cell></row><row><cell cols="3">11:35 -11:55 Alexis Block and Katherine Kuchenbecker:</cell></row><row><cell cols="3">Physical and Behavioral Factors Improve</cell></row><row><cell>Robot Hug Quality</cell><cell></cell><cell></cell></row><row><cell cols="3">11:55 -12:15 Lightning Round I: Summarized challenges</cell></row><row><cell>of contributed talks</cell><cell></cell><cell></cell></row><row><cell>12:15 -14:00 Lunch</cell><cell></cell><cell></cell></row><row><cell cols="3">14:00 -14:05 Organizers welcome to the afternoon session</cell></row><row><cell cols="3">14:05 -14:50 Invited Talk 3: Ana Paiva (University of Lis-</cell></row><row><cell cols="3">bon) Robots that listen to the users heart: The</cell></row><row><cell cols="3">role of emotions in multi-modal communica-</cell></row><row><cell cols="2">tion with social robots</cell><cell></cell></row><row><cell cols="3">14:50 -15:10 Søren Tranberg Hansen, Anders Krogsager</cell></row><row><cell cols="3">and Jakob Fredslund; A Multimodal Robot</cell></row><row><cell>Game for Seniors</cell><cell></cell><cell></cell></row><row><cell cols="3">15:10 -15:30 Lightning Round II: Summarized chal-</cell></row><row><cell cols="2">lenges of contributed talks</cell><cell></cell></row><row><cell>15:30 -15:50 Coffee Break</cell><cell></cell><cell></cell></row><row><cell>15:50 -16:20 Brainstorming</cell><cell>discussion</cell><cell>including</cell></row><row><cell cols="3">keynote speakers and presenters</cell></row><row><cell cols="3">16:20 -16:35 Presentation of brainstorming results</cell></row></table><note>(What did we learn, what are the next questions to follow) 16:35 -16:40 Summing-up and closing</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_0"> www.uc.pt   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_1"> www.deec.uc.pt   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_2"> www.isr.uc.pt   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_3">http://las.ipn.pt</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_4"> www.ipn.pt   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_5">www.miraculous-life.eu</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_6"> www.growmeup.eu   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_7"> www.growmeup.eu   </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_8">http://mrl.isr.uc.pt/projects/socialrobot</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgment</head></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This work is supported by the Horizon 2020 project GrowMeUp (Grant No. 643647) as well as the Horizon 2020 MSCA ITN ACROSSING project (Grant No. 616757).</p></div>
			</div>

			<div type="references">

				<listBibl/>
			</div>
		</back>
	</text>
</TEI>
