Carsten Trinitis’s research while affiliated with Technical University of Munich and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (123)


Advancing user-space networking for DDS message-oriented middleware: Further extensions
  • Article

February 2025

·

2 Reads

Pervasive and Mobile Computing

Vincent Bode

·

Carsten Trinitis

·

Martin Schulz

·

[...]

·

Tobias Preclik


Fig. 2: The figure illustrates that Fixed Trajectory Length (FTL) matches all experts with avoidable matching error throughout the distillation process. The figure is generated from experiments conducted on CIFAR-10 with Images Per Class (IPC) set to 1. We collect the number of cases matches with larger matching errors ∥NS − Nopt∥ ≥ γ, at every 50 iterations throughout the distillation process. The number of cases matched with larger errors fluctuates over entire process. The same can be observed with the mean value of line a. From left to right, we observe the persistence of this issue across different step size NS for FTL. Notation: #: number, N'th interval: the N'th 50 iters.
Fig. 6: The figure illustrates the influence of extended trajectory bounds, denoted as NS, on the behavior of ATT. As depicted in the first five small plots, regardless of the specific value of NS, ATT consistently exhibits a preference for selecting steps that progress from smaller to larger magnitudes. With the last small plot, we showcase ATT's selection over long distillation, where adjustments maintains. The observed behavior highlights a consistent pattern in the step selection process of ATT, emphasizing its tendency to prioritize learning from smaller steps initially before transitioning to larger ones, irrespective of the trajectory bounds under consideration.
Fig. 7: The figure demonstrate variation on ATT's test accuracy with NS. ATT's performance increasing with NS and then saturates. The experiments performed on CIFRA-10, and is in comparison with MTT.
Fig. 8: Results demonstration on CIFAR-100 with IPC=1
Performance comparison of our method ATT to other baselines on ImageNet subset. All the scores are presented in percentage. As shown, our method demonstrates significant improvement on large dataset ImageNet, especially on IPC=10

+1

Dataset Distillation by Automatic Training Trajectories
  • Preprint
  • File available

July 2024

·

107 Reads

Dataset Distillation is used to create a concise, yet informative, synthetic dataset that can replace the original dataset for training purposes. Some leading methods in this domain prioritize long-range matching, involving the unrolling of training trajectories with a fixed number of steps (NS) on the synthetic dataset to align with various expert training trajectories. However, traditional long-range matching methods possess an overfitting-like problem, the fixed step size NS forces synthetic dataset to distortedly conform seen expert training trajectories, resulting in a loss of generality-especially to those from unencountered architecture. We refer to this as the Accumulated Mismatching Problem (AMP), and propose a new approach, Automatic Training Trajectories (ATT), which dynamically and adaptively adjusts trajectory length NS to address the AMP. Our method outperforms existing methods particularly in tests involving cross-architectures. Moreover, owing to its adaptive nature, it exhibits enhanced stability in the face of parameter variations.

Download








Citations (40)


... Trajectory matching-based methods [2] optimize the synthetic dataset by aligning the parameter update trajectories of models trained on the original and synthetic data. Building on this, Du et al. [8] reduced the error in the distilled dataset by minimizing the accumulated trajectory, while Liu et al. [22] proposed an automated approach to match the closest synthetic data trajectories to the original data trajectories, improving the performance of the distilled data. These advancements collectively illustrate the progressive efforts in the field of dataset distillation to enhance efficiency, scalability, and performance. ...

Reference:

Robust Dataset Distillation by Matching Adversarial Trajectories
Dataset Distillation by Automatic Training Trajectories
  • Citing Chapter
  • November 2024

... Researchers have looked into integrating user-space networking technologies like XDP (Express Data Path) and DPDK (Data Plane Development Kit) to improve the performance of DDS middleware in order to get beyond the real-time performance limits. Bode et al. [7] showed in their study that CycloneDDS, in conjunction with DPDK, greatly lowers latency and boosts throughput, resolving some of the real-time issues with conventional DDS implementations. CycloneDDS and DPDK operate together to reduce mean latency by up to 31%, which makes it an attractive option for high-performance, networked real-time systems like largescale robotic systems or high-frequency trading. ...

Adopting User-Space Networking for DDS Message-Oriented Middleware
  • Citing Conference Paper
  • March 2024

... Using computational models to improve energy-(by pre-processing HSI images on-board satellites) and communication efficiency of CubeSats are gaining popularity, as shown by authors in [11], [12] respectively. Model compression techniques such as quantisation are used to reduce the computational complexity (and thus energy consumption) of neural network models for deployment in space applications [13], [14]. Quantised deeplearning accelerators on Intel Movidius Vision Processing Units (VPUs) have been integrated into onboard HSI imaging systems for detecting artefacts within the captured images, before transmitting them to ground stations [15]. ...

Survey of frameworks for inference of neural networks in space data systems

... This enables the satellite to downlink, for instance, only valuable non-cloudy data, thus optimizing the usage of the often limited communication channel [13]. On-board cloud segmentation can also play a role in reducing decision-making latency, increasing satellite autonomy with more optimal operations [12,13,15,25,26,27,28]. Furthermore, from a broader perspective, the models first trained for cloud binary segmentation can be retrained to solve more complex problems instead, e.g., multi-class segmentation, where each image pixel is assigned to one of various categories. ...

MACHINE LEARNING APPLICATION BENCHMARK FOR IN-ORBIT ON-BOARD DATA PROCESSING
  • Citing Conference Paper
  • June 2021

... Therefore, key performance metrics include transmission latency, CPU usage, memory consumption, and message loss rate. This paper compares the middleware used in each platform: DDS, i.e., FastDDS [6], Fig. 9. Middleware communication performance results based on data size, frequency, and number of subscribers. The first row represents data size, the second row represents frequency, and the third row represents the number of subscribers. ...

Systematic Analysis of DDS Implementations
  • Citing Conference Paper
  • November 2023

... It provides a standard interface that smoothly integrates the various system components. We chose FastDDS in this paper because of its scalability and high performance [4]. Additionally, FastDDS offers reliable, low-latency data transfer, which is very important in real-time communication. ...

DDS Implementations as Real-Time Middleware – A Systematic Evaluation
  • Citing Conference Paper
  • August 2023

... This ensures that raw data remains local while only distilled knowledge is shared. Observe that our proposal takes a different perspective from the existing Federated Distillation (FD) schemes [7,2,8]. With the objective of improving FL performance and privacy guarantees, FD distills knowledge from multiple local models independently and transfers only compact, distilled information to the server that trains a global model using this distilled data. ...

Federated Learning via Decentralized Dataset Distillation in Resource-Constrained Edge Environments
  • Citing Conference Paper
  • June 2023

... In [28], the authors integrate GPU partitioning with power-aware scheduling to optimize resource allocation in CPU-GPU heterogeneous systems under power constraints. Leveraging MIG for fine-grained workload co-location, their approach employs scalability and interference models to improve efficiency and achieve near-optimal configurations across diverse workloads. ...

Optimizing Hardware Resource Partitioning and Job Allocations on Modern GPUs under Power Caps
  • Citing Conference Paper
  • January 2023

... This enables the satellite to downlink, for instance, only valuable non-cloudy data, thus optimizing the usage of the often limited communication channel [13]. On-board cloud segmentation can also play a role in reducing decision-making latency, increasing satellite autonomy with more optimal operations [12,13,15,25,26,27,28]. Furthermore, from a broader perspective, the models first trained for cloud binary segmentation can be retrained to solve more complex problems instead, e.g., multi-class segmentation, where each image pixel is assigned to one of various categories. ...

Benchmarking and feasibility aspects of machine learning in space systems
  • Citing Conference Paper
  • May 2022

... 네트워크 기술의 발전 및 반도체 성능 향상에 따라 단위 시간당 처리되는 트랜잭션의 양도 크게 증가하는 추세이다 [1][2][3]. 특히 인공 지능을 활용한 데이터 모니터링 및 자동화된 데이터 처리가 가능해짐에 따라 이런 추세는 더욱 가속화될 것이 예상된다. 특히, 모바일 디바이스를 이용한 트랜잭션 처리나 지능화된 온라인 지불 처리 응용에서 이런 현상이 가속화되고 있다 [3] [4]. ...

Living on the Edge: Efficient Handling of Large Scale Sensor Data
  • Citing Conference Paper
  • May 2021