TY - GEN
T1 - What is the Reward for Handwriting?-A Handwriting Generation Model Based on Imitation Learning
AU - Kanda, Keisuke
AU - Iwana, Brian Kenji
AU - Uchida, Seiichi
N1 - Funding Information:
This work was partially supported by JSPS KAKENHI Grant Number JP17H06100.
Publisher Copyright:
© 2020 IEEE.
PY - 2020/9
Y1 - 2020/9
N2 - Analyzing the handwriting generation process is an important issue and has been tackled by various generation models, such as kinematics based models and stochastic models. In this study, we use a reinforcement learning (RL) framework to realize handwriting generation with the careful future planning ability. In fact, the handwriting process of human beings is also supported by their future planning ability; for example, the ability is necessary to generate a closed trajectory like '0' because any shortsighted model, such as a Markovian model, cannot generate it. For the algorithm, we employ generative adversarial imitation learning (GAIL). Typical RL algorithms require the manual definition of the reward function, which is very crucial to control the generation process. In contrast, GAIL trains the reward function along with the other modules of the framework. In other words, through GAIL, we can understand the reward of the handwriting generation process from handwriting examples. Our experimental results qualitatively and quantitatively show that the learned reward catches the trends in handwriting generation and thus GAIL is well suited for the acquisition of handwriting behavior.
AB - Analyzing the handwriting generation process is an important issue and has been tackled by various generation models, such as kinematics based models and stochastic models. In this study, we use a reinforcement learning (RL) framework to realize handwriting generation with the careful future planning ability. In fact, the handwriting process of human beings is also supported by their future planning ability; for example, the ability is necessary to generate a closed trajectory like '0' because any shortsighted model, such as a Markovian model, cannot generate it. For the algorithm, we employ generative adversarial imitation learning (GAIL). Typical RL algorithms require the manual definition of the reward function, which is very crucial to control the generation process. In contrast, GAIL trains the reward function along with the other modules of the framework. In other words, through GAIL, we can understand the reward of the handwriting generation process from handwriting examples. Our experimental results qualitatively and quantitatively show that the learned reward catches the trends in handwriting generation and thus GAIL is well suited for the acquisition of handwriting behavior.
UR - http://www.scopus.com/inward/record.url?scp=85097765456&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85097765456&partnerID=8YFLogxK
U2 - 10.1109/ICFHR2020.2020.00030
DO - 10.1109/ICFHR2020.2020.00030
M3 - Conference contribution
AN - SCOPUS:85097765456
T3 - Proceedings of International Conference on Frontiers in Handwriting Recognition, ICFHR
SP - 109
EP - 114
BT - Proceedings - 2020 17th International Conference on Frontiers in Handwriting Recognition, ICFHR 2020
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 17th International Conference on Frontiers in Handwriting Recognition, ICFHR 2020
Y2 - 7 September 2020 through 10 September 2020
ER -