TY - JOUR
T1 - 指令驱动的虚拟角色风格化动作生成
AU - Liang, Wei
AU - Huang, Yifan
AU - Shen, Yuxin
AU - Cheng, Qiyuan
N1 - Publisher Copyright:
© 2024 Beijing Institute of Technology. All rights reserved.
PY - 2024/11
Y1 - 2024/11
N2 - To solve the problems existed in instruction-driven motion generation tasks, such as inaccurate instruction comprehension and unrelated motion generation between the given task and the results, under complex scenes, a novel method was proposed based on the combination of instructions and scene information to generate the styled motions for virtual characters. Two main components were arranged in the new method, including instruction parsing and motion generation. Firstly, using a large language model, a set of limited atomic actions were predefined to parse textual instructions into subtasks with the atomic actions in the instruction parsing stage. And then, in the motion generation stage, based on a conditional variational autoencoder (cVAE) to design a frame-by-frame motion generation network, the instruction-driven styled motion generation tasks were carried out, considering various style features such as age (e.g., young, old) and style features described in the textual instructions (e.g., happy, sad). Finally, consumer investigation and four determine tests were carried out for bedroom, park, living room and cookroom scenes, validating the method efficacy, the motion authenticity and style diversity.
AB - To solve the problems existed in instruction-driven motion generation tasks, such as inaccurate instruction comprehension and unrelated motion generation between the given task and the results, under complex scenes, a novel method was proposed based on the combination of instructions and scene information to generate the styled motions for virtual characters. Two main components were arranged in the new method, including instruction parsing and motion generation. Firstly, using a large language model, a set of limited atomic actions were predefined to parse textual instructions into subtasks with the atomic actions in the instruction parsing stage. And then, in the motion generation stage, based on a conditional variational autoencoder (cVAE) to design a frame-by-frame motion generation network, the instruction-driven styled motion generation tasks were carried out, considering various style features such as age (e.g., young, old) and style features described in the textual instructions (e.g., happy, sad). Finally, consumer investigation and four determine tests were carried out for bedroom, park, living room and cookroom scenes, validating the method efficacy, the motion authenticity and style diversity.
KW - animation
KW - instruction-driven
KW - motion generation
KW - style
UR - http://www.scopus.com/inward/record.url?scp=85212583628&partnerID=8YFLogxK
U2 - 10.15918/j.tbit1001-0645.2024.017
DO - 10.15918/j.tbit1001-0645.2024.017
M3 - 文章
AN - SCOPUS:85212583628
SN - 1001-0645
VL - 44
SP - 1199
EP - 1207
JO - Beijing Ligong Daxue Xuebao/Transaction of Beijing Institute of Technology
JF - Beijing Ligong Daxue Xuebao/Transaction of Beijing Institute of Technology
IS - 11
ER -