Apple
Apple
Non-anthropomorphic Robot
Yuhan Hu Peide Huang
Apple Apple
Cupertino, United States Cupertino, United States
[email protected] [email protected]
Figure 1: Overview of our research hypothesis: robots should not only move to fulfill functional goals and constraints, i.e.,
robot moving from the initial state to goal state through a shortest, feasible trajectory (function-driven trajectory), but also use
movements to express its internal states to human counterparts during the interaction, i.e., via expression-driven trajectory to
express robot’s intention, attention, attitude, and emotions.
Abstract emotions—alongside traditional functional considerations like task
Nonverbal behaviors such as posture, gestures, and gaze are es- fulfillment, spatial constraints, and time efficiency. In this paper,
sential for conveying internal states, both consciously and un- we present the design and prototyping of a lamp-like robot that
consciously, in human interaction. For robots to interact more explores the interplay between functional and expressive objectives
naturally with humans, robot movement design should likewise in movement design. Using a research-through-design methodol-
integrate expressive qualities—such as intention, attention, and ogy, we document the hardware design process, define expressive
movement primitives, and outline a set of interaction scenario story-
Permission to make digital or hard copies of all or part of this work for personal or boards. We propose a framework that incorporates both functional
classroom use is granted without fee provided that copies are not made or distributed and expressive utilities during movement generation, and imple-
for profit or commercial advantage and that copies bear this notice and the full citation
on the first page. Copyrights for components of this work owned by others than the ment the robot behavior sequences in different function- and social-
author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or oriented tasks. Through a user study comparing expression-driven
republish, to post on servers or to redistribute to lists, requires prior specific permission
and/or a fee. Request permissions from [email protected].
versus function-driven movements across six task scenarios, our
Conference DIS’25, July 05–09, 2025, Funchal, Madeira findings indicate that expression-driven movements significantly
© 2018 Copyright held by the owner/author(s). Publication rights licensed to ACM. enhance user engagement and perceived robot qualities. This effect
ACM ISBN 978-1-4503-XXXX-X/18/06 is especially pronounced in social-oriented tasks.
https://doi.org/XXXXXXX.XXXXXXX
Conference DIS’25, July 05–09, 2025, Funchal, Madeira
Figure 4: Hardware composition of the lamp robot (left); Interaction modalities between human and robot, including gesturing,
verbal communication, light and projection display, and touch interaction (right).
also denote the trajectory 𝜏 = (𝑠 0, 𝑠 1, . . . , 𝑠𝑇 ). The reward function the goal state, such as whether it moved to the desired position,
𝑅 consists of two parts: functional utility 𝐹 and expressive utility 𝐸. turned on the light, and projected the accurate information.
Functional Utility 𝐹 defines the function-driven utility of reaching Expressive utility 𝐸, on the other hand, motivates the actions
certain states: aimed at communicating the robot’s traits, states, and intents to its
𝑇
∑︁ human interaction partners. For example, the robot may increase
𝐹 (𝜏) = 𝑓 (𝑠𝑡 ) (1) expressive utility by looking toward a book before moving to it or
𝑡 =0
displaying curiosity through head tilts. Expressive utility can be
Without loss of generality, we assume there is only one goal state
measured by users’ perceptions of the robot, including perceived
𝑠𝑔 . In this case, 𝑓 (𝑠𝑡 ) = 1 (𝑠𝑡 = 𝑠𝑔 ) , where 1 (·) is the indicator
intelligence, intuitiveness, interaction quality, trust, engagement,
function.
sense of connection, and willingness to use the robot. Drawing on
Expressive Utility 𝐸 defines the expression-oriented utility of
Theory of Mind (ToM)—the human cognitive ability to attribute
reaching certain states:
mental states such as beliefs, desires, emotions, and intentions
𝑇
∑︁ to others—we incorporate the following expression categories to
𝐸 (𝜏) = 𝑒 (𝑠𝑡 ) (2) capture expressive utilities in the design of our expressive motion
𝑡 =0
library.
In this work, we draw on design research methods to define 𝑒 (·)
along the expressive dimensions of attention, emotion, intention, Intention. Intention refers to the purpose behind the robot’s ac-
and attitudes, as discussed in Section 3.2.2. tions and the anticipation of its upcoming movements. For instance,
Finally, the objective is to maximize the total utility: when a robot extends its hand, the user can identify which object
the robot intends to pick up and what it plans to do with it, enabling
max 𝐹 (𝜏) + 𝛾𝐸 (𝜏) (3)
𝑎 0 ,...,𝑎𝑇 −1 cooperation, supervision, or intervention as needed. In the case of
where 𝛾 is the weight for the expressive utility, which could vary a lamp robot, it might briefly turn its head toward a target before
with different task and user. In Section 4, we present a user study to moving to reach or interact with it. This behavior signals the robot’s
evaluate the perception difference between the robot movements intention, indicates a shift in attention, and cues the user about the
when taking 𝛾 > 0 versus when 𝛾 = 0. next action.
3.2.2 Functional and Expressive Utility. In the context of the lamp Attention. Attention refers to where the robot’s focus is directed,
robot, Functional utility 𝐹 drives motions that aim at achieving a with gaze serving as a strong indicator of that focus. For instance,
physical goal state, such as taking the initial state of user’s reading when a robot looks at an object, it may be analyzing it or preparing
activity or an explicit verbal request, the robot moves to face the for upcoming actions. In the context of a lamp robot—where camera
book and turns the light on, as well as projects assistive information and light act as metaphoric “eyes”—we design gaze behaviors such
such as a visualization of a content in the book. The functional that looking toward the user can signal attention, for example when
utility is measured based on the level of completion of the task in the user is speaking. Similarly, a robot can exhibit joint attention by
Conference DIS’25, July 05–09, 2025, Funchal, Madeira
Figure 6: Illustration of at-home interaction scenarios, organized by the robot’s agency (proactive vs. reactive) and task context
(function-oriented vs. social-oriented).
and video prototyping. On the x-axis, we consider the primary goal skills according to the task requirements. A high-level task man-
of the human-robot interaction. In function-oriented tasks, the ager interprets the lamp’s initial placement, the environment, and
lamp robot serves as an assistant or tool—providing information contextual information to determine and activate the appropriate
displays, offering desired lighting for user activities, adapting bed- state spaces during initialization. Figure 4 (right) illustrates the var-
time lighting, and reminding users of schedules or activities. In ious modalities the robot may respond to, including user activities
contrast, social-oriented tasks position the lamp robot more like and instructive gestures, speech commands, and touch interactions.
a friend or pet, emphasizing companionship and entertainment. The robot leverages torque sensing in its joints and can potentially
Examples include suggesting creative ideas, introducing the room integrate touch sensors on its surface, enabling it to detect tactile
to visitors, engaging in playful social interactions, playing music, input and switch to compliant modes when needed.
and projecting atmospheric lighting to enhance the overall user Through the iterative design process, we selected six task scenar-
experience. ios for further implementation of function-driven and expression-
The second dimension (y-axis) reflects the robot’s agency in driven robot movements for a user study. This selection covers all
human-robot interaction, distinguishing between proactive and four sectors of the representative space, comprising three function-
reactive roles depending on the task. In robot-proactive tasks, the oriented and three social-oriented tasks, as detailed in Section 4.2.
robot initiates the interaction. Examples include sending reminders,
nudging the user to build habits, or offering creative suggestions. In
robot-reactive tasks, the robot responds to user requests or actions. 4 User Study
For instance, in a photography lighting task, the robot activates Our research question is whether movements driven by expres-
the light based on the user’s verbal instructions and adjusts its sive utility can enhance users’ perceptions of the robot and their
position in response to pointing gestures. Similarly, a sleep light experience in human-robot interaction. To investigate this, we
might switch on or off in response to the user’s movements or verbal compare two robot conditions: one employs only function-driven
commands—activating a nightlight when asked or upon detecting movements (𝛾 = 0 in Equation 3), while the other incorporates
that the user has gotten out of bed. expression-driven movements in addition to function-driven ones,
To accommodate a wide range of tasks, the robot employs mul- achieving the same goal states but through different trajectories
tiple modalities and activates different input/output channels and (𝛾 > 0). Our objective is to determine whether—and to what ex-
tent—incorporating expression design into the robot’s movements
Conference DIS’25, July 05–09, 2025, Funchal, Madeira
influences user interaction outcomes, and how these effects may the limit, and look back at the user and shake head before sending
vary according to the context of the tasks. a verbal reaction.
Remind Water. Robot interrupts use activity and sends out a
4.1 Research Questions and Hypothesis reminder to drink a cup of water. F: Move to point toward the water
RQ1: To what extent does adding expression-driven movements, cup, light up, and send a verbal reminder; E: Move to the goal pose
in addition to function-driven movements, influence users’ percep- described in F, push the cup toward the user, and gaze toward the
tions of the robot? user before sending the verbal reminder.
H1: Users will perceive a robot that combines expression-driven
and function-driven movements as more engaging, human-like, Social Conversation. Robot takes the role as a social companion,
and intelligent than one solely incorporates function-driven move- engages with the user in a social conversation about daily activ-
ments. ities. F: Respond to user’s speech verbally; E: Use movements as
RQ2: Does the task context affect movement preferences? nonverbal cues in accordance with verbal texts, including gazing
H2: Users’ perceptions will vary by task context—expression- at the user, pointing to refer to the object in context of speech, use
driven behaviors will be less favorable for function-oriented tasks kinesthetic gestures to display emotions of excitement (dancing
and more favorable for social-oriented tasks. movement) and sadness (lowering the head).
Play Music. Robot plays music entertainment accompanying
4.2 Method user’s daily activities. F: Play music with no movement; E: Play mu-
We used a within-subject study design in which each participant sic while perform dance movements, align the movement rhythm
viewed videos of the robot completing six different tasks, presented with music tempo.
in a randomized order. After watching each video, participants
rated their perception of the robot and its interaction with the 4.3 Measure
human shown in the video. They were also encouraged to explain We include six dimension of quantitative metrics to measure the
the reasoning behind their ratings, providing insights into which perception toward the robot (human-likeliness, perceived intelli-
specific robot behaviors influenced their preferences. gence, perceived emotion/character), the quality of the interaction
To create the video demonstrations, a team of human-robot in- (interaction engagement, the sense of connection), and willingness
teraction researchers and animation designers iteratively designed to use the robot in real life. Specifically, participants rate their per-
and refined pre-recorded robot movement trajectories using the de- ception on a scale from 0 to 100 to indicate their agreement to
sign primitatives proposed in section 3.2.3. These trajectories were six statements, measuring the above-mentioned aspects. Besides,
then implemented using the off-the-shelf WidowX arm controller we collected demographic data of participants including their gen-
to ensure smooth interactions. The videos used in this study are der, age, background regarding robotics, background regarding
included in the supplementary materials. expression design (such as performing arts, psychology, animation,
We designed and implemented six scenarios, each presented in communication), general level of empathy (“I find it easy to express
two conditions: empathy and understanding towards others.), general acceptance
F: A robot with function-driven movements only. toward robot (“I feel comfortable interacting with a robotic compan-
E: A robot with both function-driven and expression-driven ion”). After each video, we collect qualitative feedback of the video
movements. by asking “how would you describe the robot in the video? What do
Details of the six task scenarios and robot movement description you like or dislike about the robot?” This allow us to gather insights
are provided below. on participants reasoning of their choices and explore open-ended
Photograph Light. Robot responds to user’s hand gestures to ideas on the perception that we did not cover in the quantitative
move and offer desired lighting conditions for photography. F: Move metrics.
in response to user gesture and object position; E: Move to express
the curiosity toward the object by leaning forward, movements 4.4 Participants
incorporating robot’s attention to user command by looking back We recruited 30 participants using emails and announcements dis-
toward the user when detecting an instructive gesture. tributed within the organization1 . Responses were filtered based
on the time taken to complete the task, excluding those that took
Project Assistance. Robot observes user task and provides a cor- less than ten minutes, as well as any incomplete responses. This
responding video projection to guide the task. F: Move to a tar- process resulted in 21 valid participants (𝑁 = 21). Among them,
get position for projection, and project a corresponding video; E: eight are female, twelve are male, and one participant did not dis-
Show curiosity toward the user activity and display joint attention close their gender. The participants’ ages range from 26 to 51 years.
through gaze direction. In terms of ethnicity, ten participants self-identified as Asian, nine
self-identified as White or Caucasian, and two preferred not to
Failure Indication. User instructs a goal position for the robot
disclose their ethnicity.
which is out of reach, the robot displays the error message back to
the user. F: Attempt to move toward the goal direction, reach the 1 The study is exempt under the organization’s Human Study Review Board criteria.
limit, and verbally output the error; E: Pause to display hesitation This study fits under the research involving benign behavioral interventions and
before moving, stretch the body to display efforts when reaching collection of information from adults with their agreement (CFR 46. 104 (d) (3))
ELEGNT: Expressive and Functional Movement Design for Non-anthropomorphic Robot Conference DIS’25, July 05–09, 2025, Funchal, Madeira
Figure 7: Quantitative Results: Comparing perception ratings between expression-driven (blue) and function-driven (pink)
robot movements across six different task scenarios.
5 Results
This section presents the results through both quantitative and
qualitative lens to uncover the perception differences between the
two robot conditions across different tasks.
examine the correlations between perception metrics (average rat- matching her energy, dancing bigger because she was dancing bigger.
ings of perception) and background variables, including gender, age, It makes me want to join. There is such power in the synergy!” (P12)
general level of empathy, general acceptance of robots, backgrounds In the qualitative reasoning, participants attributed the robot with
related to robotics and character design, as shown in Figure 9. Our expressive movements characteristics of human or pets, with its
findings indicate that age significantly influenced perceptions of ex- own drives and needs. Many participants said the expressive robot
pressive robots, with older participants showing less preference for remind them of a “puppy” or “child” . In the failure indication
expressive robots (𝑝 < 0.001). Additionally, we observed a trend in scenario, P12 described the expressive robot has “a resilient spirit”.
empathy levels affecting perception differences between functional In the playing music scenario, P1 said “it looks like it’s having fun.”
and expressive robots: participants with self-rated low empathy On the contrary, participants found it difficult to attribute human
perceived a stronger increase in robot likability after the integration characteristics in robots with little movement and described the
of expressive movements. In contrast, participants who self-rated robot as a “tool” or compared it to existing home devices. Partici-
as having high empathy were less influenced by the integration of pants commented the robot with only functional movements as
expressive movement in the robot. We also found a positive corre- “boring”, “too machine-like”, “not engaging”, “emotionless”, and may
lation between robot acceptance and perception scores. However, arouse negative feelings, especially during social conversations
these correlations did not reach statistical significance. Besides, we and playing music scenarios. P3 is confused about its motivation
conducted t-tests to compare the perception difference between of asking a social-oriented question during the conversation, as
gender groups, robotic backgrounds, and groups who have or do they assumed the robot did not have its own needs or emotions.
not have backgrounds related to character and expression design, P7 also commented that the robot looked “kinda creepy as if it
including animation, psychology, performing arts, etc. Gender did was intently staring” during the social conversation due to little
not have a significant impact on perception (𝑝 = 0.2). Robotic back- movements. Participants also noted the unnaturalness and lack
ground is a significant predictor of perception, with non-roboticist of social connection with only functional movements. During the
rating robots higher than roboticist (𝑝 = 0.006). Background related scenario of reminding human to drink water, P1 commented that
to expressive character design is another strong predictor, with “It [the functional robot] does not seem to care whether or not the
experienced character designers and artists rating robots signifi- human drinks the water.” P15 mentioned that without the robot
cantly lower than others. For all the groups above, they rated the “looking at the user like how humans engage with each other, there
expressive robots higher than the functional ones. was a lack of connection.”
Even for robots with only functional movements, some par- The preferences of adding expressive movements vary across
ticipants still perceive the robot’s movement with theory of mind, tasks. For tasks that include little functional movements, and for
such as projecting robot’s attention and intention. For example, tasks that were social-oriented and less sensitive to efficiency, ex-
during the task of failure indication, the robot stretches its arm pressive movements were more appreciated. For instance, in the
toward the note before displaying the failure message, participants scenario of playing music for entertainment, P21 noted that “I really
interpreted the stretch as “the robot seemed to struggle” (P6). P7 also liked this application for robot engagement! No fast responses were
noted that the robot’s facing direction clearly indicated its attention: necessary, so having an engaging dancing motion made the robot more
“the robot and the person are looking at the same note”. engaging. ” On the other hand, for the tasks that inherently have
clear function-driven movements and are more function-oriented,
adding expressive movements could be confusing to some individu-
5.2.3 What expressive movements were valued and what
als and preferences vary. For instance, in the scenario of photograph
were not. While adding expressive movements proved to benefit
light, participants thought the expressive movements made the ro-
the interactions, some participants found them unnecessary and
bot seem less “predictable”(P5), less “steady”(P14). P18 wished the
inefficient. P15 noted for the failure indication task “... there needs
robot to “stick to only the task relevant motion which is angle and
to be a balance between engagement through motion and speed com-
the light”. Even without expressive movements, in such function-
pletion of the task being given, otherwise the human might grow
oriented tasks, participants rated highly of the robots as long as the
impatient. It might be OK the first time with the novelty factor but
robot were able to move in correspondence to the context and user
will quickly fade out.” While some participants enjoyed the expres-
request.
sive motions, others noted that some expressive behaviors could be
It is important to note that many participants were less accept-
too exaggerated, thus distracting or disturbing. Some participants
able to robot taking proactive roles than reactive roles, such as
mentioned they disliked the robot to move all the time, especially
reminding the user to drink water. For instance, P20 noted that “...
the movement for no apparent reasons, which may imply “a lack
I don’t like my life to be controlled by a robot. If I’m in the middle
of attention on the robot’s part” (P5, conversation) Most participants
of some exciting readings, I don’t want to be disrupted by a robot’s
appreciated the information that were “quick and easy to inter-
command.” Adding the expressive movements such as with a play-
pret”, while for the subtle movements, participants had different
ful character could increase the acceptance of robot behavior, P8
preferences.
noted that “Without the playfulness, I might find this type of
Participants reported negative perception when there was a mis-
interaction with a robot annoying rather than welcome and
match between robot’s movement and its perceived capabilities.
engaging.”
For the failure indication function, P14 noted that “I did not like
that it tried to get momentum with an impulse as it seems fake.” P20 5.2.4 The effect of voice and light. Participants repeatedly com-
thought the robot did not have a camera on the head thus the mented on the alignment between movements and other modalities,
“looking at notes” action seemed fake. such as the robot’s sound and light. Several participants felt there
Conference DIS’25, July 05–09, 2025, Funchal, Madeira
was a mismatch between robot’s speech cadence and the ex- interaction, increases user engagement, and can even convey an
pressive motions - while the expressive motions were “endearing, additional layer of information. On the other hand, for tasks that
showing a character”, the tone from the robot was very “automated”, are function-driven—such as adjusting lighting angles or shifting
“stiff”, and “took away from how friendly the interaction felt”. P7 and between projection spaces—adding expressive motions can disrupt
P15 noted that the timing of the voice need to align with the the robot’s primary function and potentially cause confusion or an-
timing of the movement, to make it feel more natural. P12 found noyance for users. This implies that expression-driven movements
the sound from the motors disturbing, and may only “punctuate need to complement function-driven movements by adjusting both
with smaller slower movements”. the amount and timing of expressions to enrich—rather than con-
The coordination between movement and lighting can influence flict with—the original motions. Future research should balance the
the comfort of interaction. Some participants mentioned preferring trade-off between task efficiency and characterfulness in human-
the robot to remain steady while maintaining the light. In such robot interaction, while also considering individual preferences
cases, expressive movements might interfere with the primary light- through personalized behaviors. For instance, although some users
ing function, as the robot’s motion could distract when displaying enjoyed a more animated robot, others disliked constant movement,
attention or curiosity. Additionally, the proximity of the light also particularly when it occurred without a clear or explicit reason.
impacted perceptions of disruption, as noted by P21. P7 appreciated The design and integration of expressive motions also need to
that “the robot turned out its light when looking at the person” during align with the robot’s embodiment and capabilities. For instance,
assistive projection. gaze behaviors should be co-designed with the placement of the
robot’s “eyes” (cameras) and “head”. While this may be intuitive
for humanoid robots, robots with non-anthropomorphic features
6 Discussion rely on appearance design and movement patterns to suggest a life-
In this paper, we conducted design research to explore how adding like embodiment that can be intuitive to humans and even other
expressive movements on top of purely functional ones affects species. It is equally important to match these movements with
human-robot interactions. Both quantitative and qualitative find- the robot’s other modalities—in this case, its voice and the lamp’s
ings show that expressive movements, compared to strictly func- light or display. As many participants noted, the speech content,
tional ones, enhance the overall interaction experience and improve tone, and timing during movement sequences all play a key role
perceived robot qualities. Participants were more likely to recognize in shaping the perceived quality of the robot’s behaviors. Future
the robot’s state of mind, projecting intentions, attention, emotions, research needs to consider more extensive alignment among these
and attitudes throughout their interactions. For instance, partici- different modalities to further enhance human-robot interaction.
pants recognized the robot’s “gaze” as a marker of joint attention,
suggesting a stronger bond between human and robot. Addition-
ally, participants more frequently described the expressive robot 7 Conclusion
as a living being—for example, a “pet,” a “child,” or a “friend.” In In this paper, we present ELEGNT, a framework for designing
some tasks, adding expressive movements made the experience expressive and functional movements for non-anthropomorphic
more engaging and playful. In particular, when the robot initiated robots in daily interactions. The framework integrates function-
an interaction or nudged participants, its expressive movements driven and expression-driven utilities, where the former focuses
made those interruptions feel more acceptable, such as in the case on finding an optimal path to achieve a physical goal state, and
of interrupting the user during the reading and nudging the user the latter motivates the robot to take paths that convey its internal
to drink water. This may be due to that participants possess more states—such as intention, attention, attitude, and emotion—during
empathy towards the robot with expressive movements, as they human-robot interactions. We use a lamp-shaped robot to illus-
remind them of living beings, just like the pets making the mess trate the design space for functional and expressive movements in
in the home; and thus the initially disturbing behaviors transfer various interaction scenarios, ranging from function-oriented to
into a playful social interaction. This highlighted the benefit of social-oriented tasks, and involving reactive versus proactive robot
adding expressive and characterful motions for robot-initiated task roles. We conduct a user study to compare perceptions of the robot
scenarios. when using expressive movements versus only functional move-
The quantitative results reveal differences in perception across ments across six different task scenarios. Our results indicate that
various tasks. For social-oriented tasks—such as playing music, incorporating expressive movements significantly increases user
engaging in social conversations, and nudging water—adding ex- likability toward the robot and enhances interaction engagement.
pressive movements was significantly more preferred. The qualita- The perception varies across tasks, with social-oriented tasks that
tive reasoning further illuminates this trend: in these tasks, users require minimal function-driven movements benefiting particularly
prioritize engagement and entertainment over task efficiency. Con- from the addition of expression-driven movements. Qualitative
sequently, adding expressive movements enhances the robot’s play- analysis further elaborates on users’ differing perceptions of the
fulness and character. Moreover, social-oriented tasks in the study robot’s characteristics and the perceived robots’ mental models.
generally entail minimal function-driven motions. For instance, The findings also highlight the importance of aligning movement
when playing music or holding social conversations, the robot pri- with other robot modalities, such as voice and light. Future work
marily responds verbally, and the functional output does not involve will integrate these design insights into a generative framework for
any physical movement. In these contexts, incorporating expressive creating context-aware robotic movements that effectively express
movements aligned with the social and task setting enriches the intentions in non-anthropomorphic robots.
ELEGNT: Expressive and Functional Movement Design for Non-anthropomorphic Robot Conference DIS’25, July 05–09, 2025, Funchal, Madeira