Suping Wu’s research while affiliated with Ningxia University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (32)


Towards Accurate 3D Face Alignment Under Extreme Scenarios Via Multi-Granularity Perturbation Relearning
  • Conference Paper

July 2024

·

2 Reads

Xinyu Li

·

Xing Wang

·

Xiaoxiao Yang

·

[...]

·

Xiang Zhang




Unsupervised Multi-collaborative Learning Network for 3D Face Reconstruction

January 2024

·

15 Reads

Lecture Notes in Computer Science

Monocular image-based 3D fine face reconstruction techniques aim to reconstruct 3D faces with rich face details from a single image. Existing methods have achieved remarkable results, but they cannot accurately extract light and perspective information, resulting in reconstructed faces with poor details and more noise. To this end, we propose a method for 3D face reconstruction using multi-collaborative learning network. Specifically, we design an illumination and view feature extraction network, which combines the ideas of FCN-style [8] point-by-point addition and UNet-style [7] channel dimension splicing and fusion. In this way, features at different scales can be better filtered and integrated, and key semantic information can be extracted, we can make full use of the effective features at different scales to obtain accurate light and view information. In addition, in order to be able to obtain a more comprehensive and realistic albedo characterisation, we propose a multi-resolution co-optimization module. Extensive experimental results on several evaluation datasets show that our method achieves significant improvements and excellent performance compared to state-of-the-art methods.


Self-supervised Edge Structure Learning for Multi-view Stereo and Parallel Optimization

January 2024

·

1 Read

Lecture Notes in Computer Science

Recent studies have witnessed that many self-supervised methods obtain clear progress on the multi-view stereo (MVS). However, existing methods ignore the edge structure information of the reconstructed target, which includes the outer silhouette and the edge information of the internal structure. This may lead to less satisfactory edges and completeness of the reconstruction result. To solve this problem, we propose an extractor for extracting edge structure maps, and we innovatively design an edge structure Loss to constrain the network to pay more attention to edge structure features of the reference view to improve the texture details of the reconstruction results. Specially, we utilize the idea of constructing cost volume in multi-view stereo and warp the edge structure map of the source view to the reference view to provide reliable self-supervision. In addition, we design a masking mechanism that combines local and global properties, which ensures robustness and improves the reconstruction completeness of the model for complex samples. Furthermore, we adopt an effective parallel acceleration approach to improve the training speed and reconstruction efficiency. Extensive experiments on the DTU and Tanks &Temples benchmarks demonstrate that our method improves both accuracy and completeness in comparison with other unsupervised work. In addition, our parallel method improves efficiency while ensuring accuracy. The code will be published.



A Bi-directional Optimization Network for De-obscured 3D High-Fidelity Face Reconstruction

November 2023

·

6 Reads

Communications in Computer and Information Science

3D detailed face reconstruction based on monocular images aims to reconstruct a 3D face from a single image with rich face detail. The existing methods have achieved significant results, but still suffer from inaccurate face geometry reconstruction and artifacts caused by mistaking hair for wrinkle information. To address these problems, we propose a bi-directional optimization network for de-obscured 3D high-fidelity surface reconstruction. Specifically, our network is divided into two stages: face geometry fitting and face detail optimization. In the first stage, we design a global and local bi-directional optimized feature extraction network that uses both local and global information to jointly constrain the face geometry and ultimately achieve an accurate 3D face geometry reconstruction. In the second stage, we decouple the hair and the face using a segmentation network and use the distribution of depth values in the facial region as a prior for the hair part, after which the FPU-net detail extraction network we designed is able to reconstruct finer 3D face details while removing the hair occlusion problem. With only a small number of training samples, extensive experimental results on multiple evaluation datasets show that our method achieves competitive performance and significant improvements over state-of-the-art methods.



A Detail Geometry Learning Network for High-Fidelity Face Reconstruction

September 2023

·

18 Reads

Lecture Notes in Computer Science

In this paper, we propose a Detail Geometry Learning Network (DGLN) approach to investigate the problem of self-supervised high-fidelity face reconstruction from monocular images. Unlike existing methods that rely on detail generators to generate “pseudo-details” where most of the reconstructed detail geometries are inconsistent with real faces. Our DGLN can ensure face personalization and also correctly learn more local face details. Specifically, our method includes two stages: the personalization stage and the detailization stage. In the personalization stage, we design a multi-perception interaction module (MPIM) to adaptively calibrate the weighted responses by interacting with information from different receptive fields to extract distinguishable and reliable features. To further enhance the geometric detail information, in the detailization stage, we develop a multi-resolution refinement network module (MrNet) to estimate the refined displacement map with features from different layers and different domains (i.e. coarse displacement images and RGB images). Finally, we design a novel normal smoothing loss that improves the reconstructed details and realisticity. Extensive experiments demonstrate the superiority of our method over previous work.


Citations (8)


... Three-dimensional human mesh recovery models aim to estimate the shape [1,2] and posture [3,4] Jinan Guoke Medical Technology Development Co., Ltd, Jinan, Shandong 250000, People's Republic of China mesh vertices using features extracted from images [5][6][7][8][9][10][11] or videos [12,13]. These models have various meaningful applications in fields such as virtual reality [14] and behavior analysis [15,16]. ...

Reference:

SS-MVMETRO: Semi-supervised multi-view human mesh recovery transformer
Two-stage Co-segmentation Network Based on Discriminative Representation for Recovering Human Mesh from Videos
  • Citing Conference Paper
  • June 2023

... We introduce a technique to regularize the depth map by incorporating an estimated boundary map. Our approach distinguishes itself from DEF-MVSNet [38] in terms of how edge information is represented and modeled. While DEF-MVSNet primarily focuses on determining flow directions as pixel offsets, our method explicitly learns and smooths the edge map by defining each pixel as a bimodal distribution. ...

High-Resolution Multi-View Stereo with Dynamic Depth Edge Flow
  • Citing Conference Paper
  • July 2021

... Another important method is to estimate the shape of the object and then calculate the volume of the object by multiplying a small area by the depth of the components. There are numerous methods for estimating volumes from point cloud methods [10,24,28,31]. ...

Attention-guided Multi-view Stereo Network For Depth Estimation
  • Citing Conference Paper
  • December 2020

... We quantitatively evaluate the face landmark alignment performance using the normalized mean error NME 2d (%) on the AFLW2000-3D [56] and AFLW [31] [9,13,26,33,34,36,45,47,50,53,54,56]. The corresponding results are reported in Table 1, where the lower the value, the better (the best results in each category are highlighted in bold). ...

Multi-Granularity Feature Interaction and Relation Reasoning for 3D Dense Alignment and Face Reconstruction
  • Citing Conference Paper
  • June 2021

... This paper proposes a disentangled representation learning method for single-view 3D reconstruction and face alignment. Current approaches [34] mainly decompose the face attributes and individually estimate their shape, expression, and pose parameters. Although this strategy enhances learning a single-face attribute, these methods do not consider the interaction between features. ...

Multi-Attribute Regression Network for Face Reconstruction
  • Citing Conference Paper
  • January 2021

... We quantitatively evaluate the face landmark alignment performance using the normalized mean error NME 2d (%) on the AFLW2000-3D [56] and AFLW [31] [9,13,26,33,34,36,45,47,50,53,54,56]. The corresponding results are reported in Table 1, where the lower the value, the better (the best results in each category are highlighted in bold). ...

Graph Structure Reasoning Network for Face Alignment and Reconstruction
  • Citing Chapter
  • January 2021

Lecture Notes in Computer Science

... Year SGD [37] 5.69 2021 OR-CNN [17] 3.27 2016 ARN [38] 3.25 2017 DAG-VGG16 [39] 2.81 2019 Mean-Variance Loss [40] 2.80 2018 CDCNN [22] 2.76 2018 MSFCL-KL [41] 2.73 2020 Soft-ranking [42] 2.71 2020 DEX (IMDB-WIKI) [43] 2.68 2018 CORAL-CNN [19] 2.64 2020 VDAL [44] 2.57 2020 ADPF [11] 2.54 2022 Anet+Gnet+Rnet [45] 2.47 2021 Swin Transformer [46] 2.37 2021 OURS 2.31 2024 Table 6 Performance comparison on AFAD dataset. ...

Similarity-Aware and Variational Deep Adversarial Learning for Robust Facial Age Estimation
  • Citing Article
  • January 2020

IEEE Transactions on Multimedia

... Inspired by the reinforcement learning for face-alignment tasks, in this work we propose medical landmark detection as a Markov decision process [49]. We define L = [L 1 , L 2 , · · · , L I ] ∈ R 2×I as a location vector of I points, where L i denotes for the horizontal and vertical coordinates of the i-th landmark, given a medical image I. ...

Learning Reasoning-Decision Networks for Robust Face Alignment
  • Citing Article
  • Full-text available
  • December 2018

IEEE Transactions on Pattern Analysis and Machine Intelligence