Lele Chen    

I am currently a researcher at Sony AI, focusing on addressing challenges in the realm of Generative AI. Previously, I held the role of a staff research scientist at Innopeak Technology.

I am mainly focusing on audio-visual learning, generative adversarial networks, human body and face modeling, as well as egocentric video understanding.

I hold a Ph.D. degree with the guidance of Prof. Chenliang Xu at University of Rochester, an enriching experience that spanned almost six remarkable years. In addition to this, I've had the privilege of engaging in various internships, contributing to renowned establishments like Facebook Reality Labs, Innopeak Tech ARVR Lab, visualDx Medical Image Lab and JD.com JDX Autonomous Driving Lab. These opportunities allowed me to collaborate with esteemed professionals, including Professor Zhiyao Duan, Professor Fernando De la Torre, Professor Jiebo Luo, Dr. Chen Cao, Dr. Hongda Mao, Dr. Yi Xu, Dr. Shuxue Quan, and Dr. Zhong Li.

Email  /  CV  /  Biography  /  Google Scholar  /  LinkedIn

@Vienna, May. 2015
News

I am always looking for motivated interns (remote) to join our research group in Sony AI. Please feel free to email me your resume if you are interested.

One paper about conditional image generation is conditionally accepted to Siggraph Asia 2023.

Research

Anonymous

In this manuscript, we introduce a method for modifying elements within a 3D scene based on guidance from language, masks, or bounding boxes. This innovative approach paves the way for compelling opportunities to enhance and tailor 3D objects.

Anonymous

In this paper, we propose an approach to generate high-fidelity volumetric avatar from a short monocular video.

MyStyle++: a Controllable Personalized Generative Prior

Libing Zeng, Lele Chen , Yi Xu, Nima Kalantari
Siggraph Asia 2023
paper / Supp. / bibtex / video / code / project page

In this paper, we propose an approach to obtain a personalized generative prior with explicit control over a set of attributes.

Anonymous

Luchuan Song, Lele Chen , Xiaodan Li, Zhenchao Jin, Guojun Yin, Chenliang Xu
In Submission

In this paper, we propose a new generative AI model: Artistic Video Portrait, for generating artistic portrait videos. This approach transforms a given monocular portrait video into a drivable artistic face video based on text descriptions and conditional inputs (e.g., facial expression parameters, audio) using limited data. Leveraging the spatial density function and color consistency of Neural Radiance Fields (NeRF), and the implicit features of Large Language Models (LLMs), we explore the editing capabilities for dynamic portrait videos.

Uncertainty-aware State Space Transformer for Egocentric 3D Hand Trajectory Forecasting

Wentao Bao, Lele Chen , Libing Zeng, Zhong Li, Yi Xu, Junsong Yuan, Yu Kong
ICCV 2023
paper & Supp. / bibtex / video / code / project page

In this paper, we set up an egocentric 3D hand trajectory forecasting task that aims to predict hand trajectories in a 3D space from early observed RGB videos in a first-person view.

NeuRBF: A Neural Fields Representation with Adaptive Radial Basis Functions

Zhang Chen, Zhong Li, Liangchen Song, Lele Chen , Jingyi Yu, Junsong Yuan, Yi Xu
ICCV 2023 Oral
paper / video / code / project page

We present a novel type of neural field that uses general radial bases for signal representation. State-of-the-art neural fields typically rely on grid-based representations for storing local neural features and N-dimensional linear kernels for interpolating features at continuous query points. The spatial positions of their neural features are fixed on grid nodes and cannot well adapt to target signals.

3D-aware Facial Landmark Detection via Multiview Consistent Training on Synthetic Data

Libing Zeng, Lele Chen , Wentao Bao, Zhong Li, Yi Xu, Junsong Yuan, Nima Kalantari
CVPR 2023
paper / Supp. / bibtex / video / code / project page

In this work, by leveraging synthetic data, we propose a novel multi-view consistent learning strategy to improve 3D facial landmark detection accuracy on in-the-wild images. The proposed 3D-aware module can be plugged into any learning-based landmark detection algorithm.

Anonymous

Anonymous In Submission

We propose to harness the low-frequency NeRF and leverage it to regularize the high-frequency NeRF so that it will not overfit under the few-shot setting.

Anonymous

Anonymous In Submission

In order to provide 3D shape evaluation that better aligns with human perception, we design an analytic metric named Spectrum AUC Difference (SAUCD) in this work.

NeRFPlayer: A Streamable Dynamic Scene Representation with Decomposed Neural Radiance Fields

Liangchen Song , Anpei Chen , Zhong Li , Chen Zhang, Lele Chen , Junsong Yuan, Yi Xu, Andreas Geiger
IEEE Transactions on Visualization & Computer Graphics (TVCG),
paper / bibtex / video / website

We present an efficient framework capable of fast reconstruction, compact modeling, and streamable rendering.

Real-time 3D Neural Facial Animation from Binocular Video

Chen Cao , Vasu Agrawal , Fernando De La Torre , Lele Chen , Jason Saragih , Tomas Simon , Yaser Sheikh
ACM Transactions on Graphics (SIGGRAPH) 2021
paper / bibtex / video / paper google drive link

We present a method for performing real-time facial animation of a 3D avatar from binocular video.

High-fidelity Face Tracking for AR/VR via Deep Lighting Adaptation

Lele Chen , Chen Cao , Fernando De La Torre , Jason Saragih , Chenliang Xu , Yaser Sheikh
CVPR 2021
paper / Supplymentary PDF / bibtex / video / project page

3D video avatars allows empowering virtual communications by providing compression, privacy, entertainment and a sense of presence in AR/VR. Best 3D photo-realistic AR/VR avatars driven by video, that can minimize uncanny effects, rely on person-specific models. However, existing person-specific photo-realistic 3D models are not robust to lighting and typically results in missing subtle facial behaviour, and artifacts in the avatar. This is a major drawback for the scalability of these models in communication systems (e.g., Messenger, Skype, FaceTime) and AR/VR. This paper addresses previous limitations by learning a deep learning lighting model, that in combination with a high-quality 3D face tracking algorithm provides a method for subtle and robust facial motion transfer from a regular video to a 3D photo-realistic avatar. Extensive experimental validation and comparisons to other state-of-the-art methods demonstrates the effectiveness of the proposed framework in real-world scenarios with variability in pose, expression and illumination.

Low-Latency Proactive Continuous Vision

Yiming Gan , Yuxian Qiu , Lele Chen , Jingwen Leng , Yuhao Zhu
(best paper nominee) PACT 2020
paper / bibtex / google drive link

Continuous vision is the cornerstone of a diverse range of intelligent applications found on emerging computing platforms such as autonomous machines and Augmented Reality glasses. A critical issue in today's continuous vision systems is their long end-to-end frame latency, which significantly impacts the system agility and user experience. We find that the long latency is fundamentally caused by the serialized execution model of today's continuous vision pipeline, whose key stages, including sensing, imaging, and vision computations, execute sequentially, leading to long frame latency.

Talking-head Generation with Rhythmic Head Motion

Lele Chen , Guogeng Cui , Celong Liu , Zhong Li , Ziyi Kou , Yi Xu, Chenliang Xu
ECCV 2020
paper / bibtex / code / video

Through modeling the head motion and facial expressions explicitly, manipulating 3D animation carefully, and embedding reference images dynamically, our approach achieves controllable, photo-realistic, and temporally coherent talking-head videos with natural head movements.

Example-Guided Scene Image Synthesis using Masked Spatial-Channel Attention and Patch-Based Self-Supervision
Haitian Zheng, Haofu Liao, Lele Chen, Wei Xiong, Tianlang Chen, Jiebo Luo
ECCV 2020
paper / bibtex / code

We propose to address a challenging example-guided scene image synthesis task. To propagate information between structurally uncorrelated and semantically unaligned scenes, we propose an MSCA module that leverages decoupled cross-attention for adaptive correspondence modeling. With MSCA, we propose a unified model for joint globallocal alignment and image synthesis. We further propose a patch-based self-supervision scheme that enables training. Experiments on the COCO-stuff dataset show significant improvements over the existing methods. Furthermore, our approach provides interpretability and can be extended to other content manipulation tasks.

Unsupervised Pose Flow Learning for Pose Guided Synthesis
Haitian Zheng, Lele Chen, Chenliang Xu , Jiebo Luo
IEEE Transactions on Image Processing (TIP) 2020
paper / bibtex / code

Pose-guided synthesis aims to generate a new image in an arbitrary target pose while preserving the appearance details from the source image. Existing approaches rely on either hard-coded spatial transformations or 3D body modeling. They often overlook complex non-rigid pose deformation or unmatched occluded regions, thus fail to effectively preserve appearance information. In this paper, we propose an unsupervised pose flow learning scheme that learns to transfer the appearance details from the source image. Based on such learned pose flow, we proposed GarmentNet and SynthesisNet, both of which use multi-scale feature-domain alignment for coarse-to-fine synthesis. Experiments on the DeepFashion, MVC dataset and additional real-world datasets demonstrate that our approach compares favorably with the state-of-the-art methods and generalizes to unseen poses and clothing styles.

3D Human Avatar Digitization from a Single Image
Zhong Li *, Lele Chen *, Celong Liu , Yu Gao, Yuanzhou Ha , Chenliang Xu , Shuxue Quan, Yi Xu (* equal contribution)
(best paper award) VRCAI, 2019, Computer & Graphics (C & G), 2021
paper / bibtex / google drive link

In this paper, we propose a pipeline that reconstructs a 3D human shape avatar at a glance. Our approach simultaneously reconstructs the three-dimensional human geometry and whole body texture map with only a single RGB image as input.

TailorGAN: Making User-Defined Fashion Designs
Lele Chen, Guo Li, Justin Tian, Cheng-Haw Wu, Erh-Kan King, Kuan-Ting Chen, Shao-Hang Hsieh, Chenliang Xu
WACV, 2020
paper / code

We consider a task: given a reference garment image A and another image B with target attribute (collar/sleeve), generate a photo-realistic image which combines the texture from reference A and the new attribute from reference B.

Hierarchical Cross-modal Talking Face Generation with Dynamic Pixel-wise Loss
Lele Chen , Ross K. Maddox , Zhiyao Duan , Chenliang Xu
CVPR 2019
paper / video / bibtex / code / project

We devise a cascade GAN approach to generate a talking face video, which is robust to different face shapes, view angles, facial characteristics, and noisy audio conditions. Instead of learning a direct mapping from audio to video frames, we propose first to transfer audio to high-level structure, i.e., the facial landmarks, and then to generate video frames conditioned on the landmarks. Compared to a direct audio-to-image approach, our cascade approach avoids fitting spurious correlations between audiovisual signals that are irrelevant to the speech content. We, humans, are sensitive to temporal discontinuities and subtle artifacts in video. To avoid those pixel jittering problems and to enforce the network to focus on audiovisual-correlated regions, we propose a novel dynamically adjustable pixel-wise loss with an attention mechanism.

Lip Movements Generation at a Glance
Lele Chen*, Zhiheng Li*, Ross K Maddox, Zhiyao Duan, Chenliang Xu (* equal contribution)
ECCV 2018
paper / poster / video / bibtex / news / code / project

Given an arbitrary audio speech and one lip image of an arbitrary target identity, generate synthesized lip movements of the target identity saying the speech. To perform well, a model needs to not only consider the retention of target identity, photo-realistic of synthesized images, consistency and smoothness of lip images in a sequence, but more importantly, learn the correlations between audio speech and lip movements.

Deep cross-modal audio-visual generation
Lele Chen, Sudhanshu Srivastava, Zhiyao Duan, Chenliang Xu
ACM MMW 2017
paper / poster / bibtex / code / Sub-URMP dataset / URMP dataset

We have developed algorithms in audio-visual source association that are able to segment corresponding audio-visual data pairs; we have created deep generative neural networks utilizing adversarial training that are able to generate one modality, i.e., audio/visual, from the other modality, i.e., visual/audio. The outputs of cross-modal generation are beneficial to many applications, such as aiding hearing- or visually-impaired and content creation in virtual reality.

MRI Tumor Segmentation with Densely Connected 3D CNN
Lele Chen, Yue Wu, Adora M. DSouza , Anas Z. Abidin, Axel Wismuller, Chenliang Xu
SPIE Image Processing 2017
paper / slides / bibtex / code

In this paper, we introduce a new approach for brain tumor segmentation in MRI scans. DenseNet was initially introduced for the image classification problem. In this work, we explore the potential of densely connected blocks in 3D segmentation tasks. Compared with traditional networks with no skip connections, the improved information flow extracts better features and significantly help the optimization. We take multi-scale receptive fields into account to accurately classify voxels.

Toward a visual assistive listening device: Real-time synthesis of a virtual talking face from acoustic speech using deep neural networks
Lele Chen, Emre Eskimez, Zhiheng Li, Zhiyao Duan , Chenliang Xu , Ross K Maddox
The Journal of the Acoustical Society of America 2018
paper / bibtex

Speech perception is a crucial function of the human auditory system, but speech is not only an acoustic signal-visual cues from a talker's face and articulators (lips, teeth, and tongue) carry considerable linguistic information. These cues offer substantial and important improvements to speech comprehension when the acoustic signal suffers degradations like background noise or impaired hearing. However, useful visual cues are not always available, such as when talking on the phone or listening to a podcast. We are developing a system for generating a realistic speaking face from speech audio input. The system uses novel deep neural networks trained on a large audio-visual speech corpus. It is designed to run in real time so that it can be used as an assistive listening device. Previous systems have shown improvements in speech perception only for the most degraded speech. Our design differs notably from earlier ones in that it does not use a language model-instead, it makes a direct transformation from speech audio to face video. This allows the temporal coherence between the acoustic and visual modalities to be preserved, which has been shown to be crucial to cross-modal perceptual binding.

Professional Activities

I am a reviewer of IEEE TIP / IEEE TMM / IEEE Access / Neurocomputing / WACV / CVPR / ICCV / AAAI / ECCV / IEEE-TPAMI / ACM Siggraph, etc.

Teaching

CSC 577 Advanced Topics in Computer Vision- Fall 2019

This course covers advanced research topics in computer vision with an emphasis on learning structured representations and embeddings. Approaches for learning from unimodal (e.g., images and videos), and multimodal data (e.g., vision and language, vision and audio) will be covered and include topics from structured predications, deep learning and others. The course will be a mix of lecture, student presentation and discussion. Prerequisites: CSC 249/449 or CSC 246/446 or CSC 298/578 (Deep Learning and Graphical Models) or permission of the instructor.


CIS 418 Advanced Business Modeling & Analysis- Spring 2019

This course covers a range of skills and methodologies for the use of spreadsheet-based tools in business modeling and quantitative analysis. We will also develop some introductory techniques for quantitative analysis in R (no prior R programming experience is required). Topics include forecasting using time series regression analysis and ARIMA models; clustering and classification models; linear optimization models and sensitivity analysis; and Monte Carlo simulation. The models and methods covered in CIS 418 can be used across diverse applications in finance, economics, marketing, logistics, and operational management.


CIS 442 Data Management for Analytics- Winter 2018

Companies and organizations in every industry rely heavily on their data, not only for the day-to-day tactical needs of running the business but also to generate information and analysis to support problem solving and decision-making in a wide variety of applications. Database design and management along with systems/techniques for data retrieval, transformation and analysis have the potential to play a key role in business strategy and to create business value and competitive advantage.


CSC 461 Database Systems- Winter 2018

This course presents the fundamental concepts of database design and use. It provides a study of data models, data description languages, and query facilities including relational algebra and SQL, data normalization, transactions and their properties, physical data organization and indexing, security issues and object databases. It also looks at the new trends in databases, for example, Big Data, MapReduce, and NoSQL. The knowledge of the above topics will be applied in the design and implementation of a database application using a target database management system as part of a semester-long group project.


CIS 442F Big Data - Spring 2018

This class offers an introduction to big data concepts, environments, processes, and tools from the perspective of data analysts and data scientists. The course will set the scene for the emergence of big data as an important trend in the business world and explain the technical architectures that make analyzing data at scale possible. The hands-on portion of the class will focus on the major tools of the Hadoop big data ecosystem such as HDFS, Pig, Hive, Sqoop, Hue, Zeppelin, and Spark. In addition, students will gain a broad understanding of the role of MapReduce, Tez, Impala, YARN, and other big data technologies.


GBA 464 Programming for Analytics - Fall 2017

This course provides some foundations for programming in the R environment. We cover traditional programming concepts such as operators, data structures, control structures, repetition and user-defined functions. In addition, these concepts will be taught in the context of marketing and business analytics problems related to data management and visualization. Other than high-level programming, the students will gain a foundational understanding of how data is can be stored, organized and pulled, in given data analytics context.


Other

CSC400 Graduate Problem Seminar - Fall 2018

You can find my NSF ITRG Miniproposal in here .



UP HOME
URCS People | URCS Home Page

this guy's website is awesome