MMSP-P2.3
TALKING FACE GENERATION FOR IMPRESSION CONVERSION CONSIDERING SPEECH SEMANTICS
Saki Mizuno, Nobukatsu Hojo, Kazutoshi Shinoda, Keita Suzuki, Mana Ihori, Hiroshi Sato, Tomohiro Tanaka, Naotaka Kawata, Satoshi Kobashikawa, Ryo Masumura, NTT Corporation, Japan
Session:
MMSP-P2: Multimedia Generation and Synthesis Poster
Track:
Multimedia Signal Processing
Location:
Poster Zone 5A
Poster Board PZ-5A.3
Poster Board PZ-5A.3
Presentation Time:
Tue, 16 Apr, 16:30 - 18:30 (UTC +9)
Session Chair:
Jongyoo Kim, Yonsei University, South Korea
Session MMSP-P2
MMSP-P2.1: ENHANCING REALISM IN 3D FACIAL ANIMATION USING CONFORMER-BASED GENERATION AND AUTOMATED POST-PROCESSING
Yi Zhao, Chunyu Qiang, Hao Li, Kuaishou, China; Yulan Hu, Renmin University of China, China; Wangjin Zhou, Kyoto University, Japan; Sheng Li, NICT, Japan, Japan
MMSP-P2.2: SimFall: A Data Generator For RF-based Fall Detection
Jiamu Li, Dongheng Zhang, Qi Chen, Yadong Li, Jianyang Wang, Wenxuan Li, Yang Hu, Qibin Sun, Yan Chen, University of Science and Technology of China, China
MMSP-P2.3: TALKING FACE GENERATION FOR IMPRESSION CONVERSION CONSIDERING SPEECH SEMANTICS
Saki Mizuno, Nobukatsu Hojo, Kazutoshi Shinoda, Keita Suzuki, Mana Ihori, Hiroshi Sato, Tomohiro Tanaka, Naotaka Kawata, Satoshi Kobashikawa, Ryo Masumura, NTT Corporation, Japan
MMSP-P2.4: Visually Guided Binaural Audio Generation with Cross-modal Consistency
Miao Liu, Jing Wang, Beijing Institute of Technology, China; Xinyuan Qian, University of Science and Technology Beijing, China; Xiang Xie, Beijing Institute of Technology, China
MMSP-P2.5: BINAURALMUSIC: A DIVERSE DATASET FOR IMPROVING CROSS-MODAL BINAURAL AUDIO GENERATION
Yunqi Li, School of Data Science and Intelligent Media, Communication University of China, Beijing 100024, China, China; Shulin Liu, School of Information and Communication Engineering, Communication University of China, Beijing 100024, China, China; Haonan Cheng, Long Ye, State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing 100024, China, China
MMSP-P2.6: ENHANCING SPATIAL AUDIO GENERATION WITH SOURCE SEPARATION AND CHANNEL PANNING LOSS
Wootaek Lim, KAIST, ETRI, Korea, Republic of; Juhan Nam, KAIST, Korea, Republic of
MMSP-P2.7: TEXT-DRIVEN TALKING FACE SYNTHESIS BY REPROGRAMMING AUDIO-DRIVEN MODELS
Jeongsoo Choi, Minsu Kim, Se Jin Park, Yong Man Ro, Korea Advanced Institute of Science and Technology, Korea, Republic of
MMSP-P2.8: EMOTALKER: EMOTIONALLY EDITABLE TALKING FACE GENERATION VIA DIFFUSION MODEL
Bingyuan Zhang, Xulong Zhang, Ning Cheng, Ping An Technology (Shenzhen) Co., Ltd., China; Jun Yu, University of Science and Technology of China, Hefei, China, China; Jing Xiao, Jianzong Wang, Ping An Technology (Shenzhen) Co., Ltd., China
MMSP-P2.9: SPEECH-DRIVEN EMOTIONAL 3D TALKING FACE ANIMATION USING EMOTIONAL EMBEDDINGS
Seongmin Lee, Jeonghaeng Lee, Hyewon Song, Sanghoon Lee, Yonsei University, Korea, Republic of
MMSP-P2.10: PHISANET: PHONETICALLY INFORMED SPEECH ANIMATION NETWORK
Salvador Medina, Sarah Taylor, Carsten Stoll, Epic Games, United States of America; Gareth Edwards, Cubic Motion, United Kingdom of Great Britain and Northern Ireland; Alex Hauptmann, Shinji Watanabe, Carnegie Mellon University, United States of America; Iain Matthews, Epic Games, United States of America
MMSP-P2.11: ENHANCING IMAGE-TEXT MATCHING WITH ADAPTIVE FEATURE AGGREGATION
Zuhui Wang, Yunting Yin, I.V. Ramakrishnan, State University of New York at Stony Brook, United States of America
Contacts