TY - JOUR
T1 - User modelling using multimodal information for personalised dressing assistance
AU - Gao, Yixing
AU - Chang, Hyung Jin
AU - Demiris, Yiannis
PY - 2020/3/20
Y1 - 2020/3/20
N2 - Assistive robots in home environments are steadily increasing in popularity. Due to significant variabilities in human behaviour, as well as physical characteristics and individual preferences, personalising assistance poses a challenging problem. In this paper, we focus on an assistive dressing task that involves physical contact with a human’s upper body, in which the goal is to improve the comfort level of the individual. Two aspects are considered to be significant in improving a user’s comfort level: having more natural postures and exerting less effort. However, a dressing path that fulfils these two criteria may not be found at one time. Therefore, we propose a user modelling method that combines vision and force data to enable the robot to search for an optimised dressing path for each user and improve as the human-robot interaction progresses. We compare the proposed method against two single-modality state-of-the-art user modelling methods designed for personalised assistive dressing by user studies (31 subjects). Experimental results show that the proposed method provides personalised assistance that results in more natural postures and less effort for human users.
AB - Assistive robots in home environments are steadily increasing in popularity. Due to significant variabilities in human behaviour, as well as physical characteristics and individual preferences, personalising assistance poses a challenging problem. In this paper, we focus on an assistive dressing task that involves physical contact with a human’s upper body, in which the goal is to improve the comfort level of the individual. Two aspects are considered to be significant in improving a user’s comfort level: having more natural postures and exerting less effort. However, a dressing path that fulfils these two criteria may not be found at one time. Therefore, we propose a user modelling method that combines vision and force data to enable the robot to search for an optimised dressing path for each user and improve as the human-robot interaction progresses. We compare the proposed method against two single-modality state-of-the-art user modelling methods designed for personalised assistive dressing by user studies (31 subjects). Experimental results show that the proposed method provides personalised assistance that results in more natural postures and less effort for human users.
KW - Multimodal user modelling
KW - assistive dressing
KW - human-robot interaction
KW - vision and force fusion
UR - http://www.scopus.com/inward/record.url?scp=85082066456&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2020.2978207
DO - 10.1109/ACCESS.2020.2978207
M3 - Article
SN - 2169-3536
VL - 8
SP - 45700
EP - 45714
JO - IEEE Access
JF - IEEE Access
M1 - 9024050
ER -