Tianyu Shen’s research while affiliated with Beijing University of Chemical Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (7)


Fixed-Time Composite Learning Fuzzy Control With Disturbance Rejection for Uncertain Engineering Systems Toward Industry 5.0
  • Article

July 2024

·

12 Reads

·

1 Citation

IEEE Transactions on Systems Man and Cybernetics Systems

·

Yi Zhang

·

Yafei Chang

·

[...]

·

Intelligent control is a crucial technology for realizing Industry 5.0, which makes industrial engineering systems more efficient, robust, and resilient. It is noteworthy that uncertainties and disturbances will inevitably be detrimental to the control performances of Industry 5.0 engineering applications. To deal with these issues, we propose a novel super-twisting-like continuous extended state observer-based fixed-time composite learning fuzzy control scheme and apply it to a typical engineering system. Unlike conventional fixed-time adaptive fuzzy control methods that update parameters merely by closed-loop stability conditions, the proposed fixed-time control scheme utilizes both tracking errors and prediction errors to update parameters compositely, which achieves better-tracking performance and fuzzy approximation precision. First, fuzzy logic systems (FLSs) are developed to identify the unknown model functions in the Industry 5.0 engineering system. Second, to deal with the remaining approximation errors of the FLSs, parameter uncertainties, and external disturbances, the novel super-twisting-like continuous extended state observers are designed to estimate these lumped disturbances. Third, the prediction errors that indicate the fuzzy approximation precision are constructed by developing fixed-time parallel estimators. Moreover, rigorous Lyapunov stability analysis is carried out to illustrate the fixed-time convergence of the entire closed-loop control system. Finally, the proposed control scheme is applied to a practical buck converter engineering system toward Industry 5.0, and comparative hardware experiments verified the advantages of the control scheme.


The Journey/DAO/TAO of Embodied Intelligence: From Large Models to Foundation Intelligence and Parallel Intelligence

June 2024

·

22 Reads

·

3 Citations

IEEE/CAA Journal of Automatica Sinica

The tremendous impact of large models represented by ChatGPT [1]–[3] makes it necessary to consider the practical applications of such models [4]. However, for an artificial intelligence (AI) to truly evolve, it needs to possess a physical “body” to transition from the virtual world to the real world and evolve through interaction with the real environments. In this context, “embodied intelligence” has sparked a new wave of research and technology, leading AI beyond the digital realm into a new paradigm that can actively act and perceive in a physical environment through tangible entities such as robots and automated devices [5].


ChatGPT-Based Scenario Engineer: A New Framework on Scenario Generation for Trajectory Prediction

March 2024

·

63 Reads

·

14 Citations

IEEE Transactions on Intelligent Vehicles

The latest developments in parallel driving foreshadow the possibility of delivering intelligence across organizations using foundation models. As is well-known, there are limitations in scenario acquisition in the field of intelligent vehicles (IV), such as efficiency, diversity, and complexity, which hinder in-depth research of vehicle intelligence. To address this issue, this manuscript draws inspiration from scenarios engineering, parallel driving and introduces a pioneering framework for scenario generation, leveraging the ChatGPT, denoted as SeGPT. Within this framework, we define a trajectory scenario and design prompts engineering to generate complex and challenging scenarios. Furthermore, SeGPT, in combination with “Three Modes”, foundation models, vehicle operating system, and other advanced infrastructure, holds the potential to achieve higher levels of autonomous driving. Experimental outcomes substantiate SeGPT's adeptness in producing a spectrum of varied scenarios, underscoring its potential to augment the development of trajectory prediction algorithms. These findings mark significant progress in the domain of scenario generation, also pointing towards new directions in the research of vehicle intelligence and scenarios engineering.



Parallel Vision for Long-Tail Regularization: Initial Results From IVFC Autonomous Driving Testing

June 2022

·

38 Reads

·

80 Citations

IEEE Transactions on Intelligent Vehicles

Long-tail effect, characterized by highly frequent occurrence of normal scenarios and the scarce appearance of extreme “long-tail” scenarios, ubiquitously exists in the vision-related problems in the real-world applications. Though many computer vision methods to date have already achieved feasible performance for most of the normal scenarios, it is still challenging for existing vision systems to accurately perceive the long-tail scenarios. This deficiency largely hinders the practical application of computer vision systems, since long-tail problems may incur fatal consequences, such as traffic accidents, taking the vision systems of autonomous vehicles as an example. In this paper, we firstly propose a theoretical framework named Long-tail Regularization (LoTR), for analyzing and tackling the long-tail problems in the vision perception of autonomous driving. LoTR is able to regularize the scarcely occurred long-tail scenarios to be frequently encountered. Then we present a Parallel Vision Actualization System (PVAS), which consists of closed-loop optimization and virtual-real interaction, to search for challenging long-tail scenarios and produce large-scale long-tail driving scenarios for autonomous vehicles. In addition, we introduce how to perform PVAS in Intelligent Vehicle Future Challenge of China (IVFC), the most durable autonomous driving competition around the world. Results over the past decade demonstrate that PVAS can effectively guide the collection of long-tail data to diminish the cost in the real world, and thus promote the capability of vision systems to adapt to complex environments, alleviating the impact of long-tail effect.


Depth-Aware Multi-Person 3D Pose Estimation With Multi-Scale Waterfall Representations

January 2022

·

9 Reads

·

8 Citations

IEEE Transactions on Multimedia

Estimating absolute 3D poses of multiple people from monocular image is challenging due to the presence of occlusions and the scale variation among different persons. Among the existing methods, the top-down paradigms are highly dependent on human detection which is prone to the influence from inter-person occlusions, while the bottom-up paradigms suffer from the difficulties in keypoint feature extraction caused by scale variation and unreliable joint grouping caused by occlusions. To address these challenges, we introduce a novel multi-person 3D pose estimation framework, aided by multi-scale feature representations and human depth perceiving. Firstly, a waterfall-based architecture is incorporated for multi-scale feature representations to achieve a more accurate estimation of occluded joints with a better detection of human shapes. Then the global and local representations are fused for handling the effects of inter-person occlusion and scale variation in depth perceiving and keypoint feature extraction. Finally, with the guidance of the fused multi-scale representations, a depth-aware model is exploited for better 2D joint grouping and 3D pose recovering. Quantitative and qualitative evaluations on benchmark datasets of MuCo-3DHP and MuPoTS-3D prove the effectiveness of our proposed method. Furthermore, we produce an occluded MuPoTS-3D dataset and the experiments on it validate the superiority of our method for overcoming the occlusions.


Fig. 1. Layered architecture of CPSS for Edu-Metaverse in cyber-physicalsocial space.
Fig. 3. (a) Implementation example of the camera setup in the physical-world classroom with the V-Classroom's local coordinate system. (b) Color images of the physical-world classroom with learners captured by the three cameras, and coordinate transformations in the V-Classroom system for camera calibration and novel view synthesis.
VirtualClassroom: A Lecturer-Centered Consumer-Grade Immersive Teaching System in Cyber–Physical–Social Space
  • Article
  • Full-text available

January 2022

·

90 Reads

·

15 Citations

IEEE Transactions on Systems Man and Cybernetics Systems

Lecturers, as the guidance of the classroom, play a significant role in the teaching process. However, the lecturers’ sense of space immersion has been ignored in current virtual teaching systems. In this article, we explore the cyber–physical–social intelligence for Edu-Metaverse in cyber–physical–social space and specially design a lecturer-centered immersive teaching system, taking the social and lecturers’ factors into consideration. We call this system VirtualClassroom (V-Classroom). Specifically, we first introduce the cyber–physical–social system (CPSS) paradigm of V-Classroom so that the workflow is standardized and significantly simplified, and the systems can be constructed with off-the-shelf hardware. The key component of V-Classroom is a cyber-world representation of a physical-world classroom instrumented with sparse consumer-grade RGBD cameras for capturing the 3-D geometry and texture of the classrooms. We provide each V-Classroom lecturer with a physical device for sending 6DoF view-change messages and showing view-dependent content of the remote classroom. Following the above paradigm, we develop the V-Classroom algorithms, including V-Classroom depth algorithm (V-DA) and V-Classroom view algorithm (V-VA), to achieve the real-time rendering of remote classrooms. V-DA is dedicated to recovering accurate depth information of the classrooms while V-VA is devoted to real-time novel view synthesis. Finally, we illustrate our implemented CPSS-driven V-Classroom prototype, based on real-world classroom scenarios we collected, and discuss the main challenges and future direction.

Download

Citations (6)


... T He evolution of autonomous driving technology heralds a transformative era in transportation systems, where the depth and integration of environmental perception are central to enhancing road safety, easing traffic pressures, and optimizing energy utilization [1]. While autonomous driving has made commendable strides towards fully autonomous operations, its perceptual limitations continue to pose significant Manuscript challenges for its broader adoption [2]. ...

Reference:

MIPD: A Multi-sensory Interactive Perception Dataset for Embodied Intelligent Driving
The Journey/DAO/TAO of Embodied Intelligence: From Large Models to Foundation Intelligence and Parallel Intelligence
  • Citing Article
  • June 2024

IEEE/CAA Journal of Automatica Sinica

... However, leveraging LLMs for scenario generation remains relatively unexplored. Li et al. (2024b) introduced a framework utilizing ChatGPT to generate trajectory data, but its raw outputs lack physical realism and consistency with realworld scenarios, requiring active user intervention through Chain-of-Thought prompting (Zhang et al. 2022). TARGET (Deng et al. 2023) have utilized LLMs for generating simulator scenarios by incorporating a domain-specific language to create configurations based on self-defined traffic rules. ...

ChatGPT-Based Scenario Engineer: A New Framework on Scenario Generation for Trajectory Prediction
  • Citing Article
  • March 2024

IEEE Transactions on Intelligent Vehicles

... The success of NeRF (Mildenhall et al. 2021) and subsequent works (Trevithick and Yang 2021;Wang et al. 2021b;Yu et al. 2021) have achieved impressive novel view synthesis applications. To overcome the drawback of dense input views, multiple works propose to extra regularizations or priors for sparse view novel view synthesis, such as depth and appearance smoothness (RegNeRF (Niemeyer et al. 2022), MVSNeRF ), ray entropy regularization (InfoNeRF (Kim, Seo, and Han 2022)), perceptual losses (SVS (González et al. 2022)), Spatio-Temporal consistency (Li et al. 2023) or ray distortion (Mip-NeRf360 ) et al. Besides, some recent approaches (Wei et al. 2021;Deng et al. 2022;Roessle et al. 2022) use depth priors to constrain the NeRF optimization, which also achieves promising novel view synthesis results from sparse input views. ...

Dynamic View Synthesis with Spatio-Temporal Feature Warping from Sparse Views
  • Citing Conference Paper
  • October 2023

... From the perspective of model methods, existing models can be classified into direct estimation and 2D-to-3D lifting methods. Direct estimation methods [9][10][11][12] directly estimate 3D poses from input images or videos without requiring additional intermediate steps to estimate corresponding 2D pose coordinates. Conversely, 2D-to-3D lifting methods [17-21,] rely on obtained 2D poses to estimate corresponding 3D pose coordinates. ...

Depth-Aware Multi-Person 3D Pose Estimation With Multi-Scale Waterfall Representations
  • Citing Article
  • January 2022

IEEE Transactions on Multimedia

... Virtual spaces can be, for example, single immersive virtual environments (IVE) provided by different Metaverse platforms [3], [8], [9], [10], [12], [13], [15], [16] or complementary services (e.g., learning management systems) [5], [8]. While the Metaverse is not inherently dependent on virtual reality (VR), VR holds a pivotal role as a core technology [4], [6], [10], [15], [17], [18], providing IVE. The vision of the Metaverse is that IVE can be generated and used by programmers, everyday people, commercial or political users. ...

VirtualClassroom: A Lecturer-Centered Consumer-Grade Immersive Teaching System in Cyber–Physical–Social Space

IEEE Transactions on Systems Man and Cybernetics Systems

... Safety-critical scenarios are crucial for diagnosing functional defects in autonomous vehicles. Acquiring these random scenarios directly from natural driving data is highly inefficient [4]. Therefore, identifying and discovering safety-critical scenarios is a complex and challenging task. ...

Parallel Vision for Long-Tail Regularization: Initial Results From IVFC Autonomous Driving Testing
  • Citing Article
  • June 2022

IEEE Transactions on Intelligent Vehicles