George Kesidis’s research while affiliated with York College of PA and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (380)


XRFlux: Virtual Reality Benchmark for Edge Caching Systems
  • Preprint

December 2024

·

George Kesidis

We introduce a Unity based benchmark XRFlux for evaluating Virtual Reality (VR) delivery systems using edge-cloud caching. As VR applications and systems progress, the need to meet strict latency and Quality of Experience (QoE) requirements is increasingly evident. In the context of VR, traditional cloud architectures (e.g., remote AWS S3 for content delivery) often struggle to meet these demands, especially for users of the same application in different locations. With edge computing, resources are brought closer to users in efforts to reduce latency and improve QoEs. However, VR's dynamic nature, with changing fields of view (FoVs) and user synchronization requirements, creates various challenges for edge caching. We address the lack of suitable benchmarks and propose a framework that simulates multiuser VR scenarios while logging users' interaction with objects within their actual and predicted FoVs. The benchmark's activity log can then be played back through an edge cache to assess the resulting QoEs. This tool fills a gap by supporting research in the optimization of edge caching (and other edge-cloud functions) for VR streaming.


Ten Ways in which Virtual Reality Differs from Video Streaming

November 2024

·

1 Read

Virtual Reality (VR) applications have a number of unique characteristics that set them apart from traditional video streaming. These characteristics have major implications on the design of VR rendering, adaptation, prefetching, caching, and transport mechanisms. This paper contrasts VR to video streaming, stored 2D video streaming in particular, and discusses how to rethink system and network support for VR.




BIC-Based Mixture Model Defense Against Data Poisoning Attacks on Classifiers: A Comprehensive Study

August 2024

·

6 Reads

·

1 Citation

IEEE Transactions on Knowledge and Data Engineering

Data Poisoning (DP) is an effective attack that causes trained classifiers to misclassify their inputs. DP attacks significantly degrade a classifier's accuracy by covertly injecting attack samples into the training set. Broadly applicable to different classifier structures, without strong assumptions about the attacker, an unsupervised Bayesian Information Criterion (BIC)-based mixture model defense against “error generic” DP attacks is herein proposed that: 1) addresses the most challenging embedded DP scenario wherein, if DP is present, the poisoned samples are an a priori unknown subset of the training set, and with no clean validation set available; 2) applies a mixture model both to well-fit potentially multi-modal class distributions and to capture poisoned samples within a small subset of the mixture components; 3) jointly identifies poisoned components and samples by minimizing the BIC cost defined over the whole training set, with the identified poisoned data removed prior to classifier training. Our experimental results, for various classifier structures and benchmark datasets, demonstrate the effectiveness of our defense under strong DP attacks, as well as its superiority over other DP defenses.



On Trojans in Refined Language Models

June 2024

·

3 Reads

Backdoor data poisoning, inserted within instruction examples used to fine-tune a foundation Large Language Model (LLM) for downstream tasks (\textit{e.g.,} sentiment prediction), is a serious security concern due to the evasive nature of such attacks. The poisoning is usually in the form of a (seemingly innocuous) trigger word or phrase inserted into a very small fraction of the fine-tuning samples from a target class. Such backdoor attacks can: alter response sentiment, violate censorship, over-refuse (invoke censorship for legitimate queries), inject false content, or trigger nonsense responses (hallucinations). In this work we investigate the efficacy of instruction fine-tuning backdoor attacks as attack "hyperparameters" are varied under a variety of scenarios, considering: the trigger location in the poisoned examples; robustness to change in the trigger location, partial triggers, and synonym substitutions at test time; attack transfer from one (fine-tuning) domain to a related test domain; and clean-label vs. dirty-label poisoning. Based on our observations, we propose and evaluate two defenses against these attacks: i) a \textit{during-fine-tuning defense} based on word-frequency counts that assumes the (possibly poisoned) fine-tuning dataset is available and identifies the backdoor trigger tokens; and ii) a \textit{post-fine-tuning defense} based on downstream clean fine-tuning of the backdoored LLM with a small defense dataset. Finally, we provide a brief survey of related work on backdoor attacks and defenses.



Temporal-Distributed Backdoor Attack against Video Based Action Recognition

March 2024

·

2 Reads

·

3 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Deep neural networks (DNNs) have achieved tremendous success in various applications including video action recognition, yet remain vulnerable to backdoor attacks (Trojans). The backdoor-compromised model will mis-classify to the target class chosen by the attacker when a test instance (from a non-target class) is embedded with a specific trigger, while maintaining high accuracy on attack-free instances. Although there are extensive studies on backdoor attacks against image data, the susceptibility of video-based systems under backdoor attacks remains largely unexplored. Current studies are direct extensions of approaches proposed for image data, e.g., the triggers are independently embedded within the frames, which tend to be detectable by existing defenses. In this paper, we introduce a simple yet effective backdoor attack against video data. Our proposed attack, adding perturbations in a transformed domain, plants an imperceptible, temporally distributed trigger across the video frames, and is shown to be resilient to existing defensive strategies. The effectiveness of the proposed attack is demonstrated by extensive experiments with various well-known models on two video recognition benchmarks, UCF101 and HMDB51, and a sign language recognition benchmark, Greek Sign Language (GSL) dataset. We delve into the impact of several influential factors on our proposed attack and identify an intriguing effect termed "collateral damage" through extensive studies.



Citations (51)


... When the accuracy of backdoored samples drops to 0%, the model accuracy on genuine unlearned data and test data is still around 80%. These results are consistent with current backdoor studies [11,33,42]. We present more detailed discussion in Appendix A. ...

Reference:

TAPE: Tailored Posterior Difference for Auditing of Machine Unlearning
MM-BD: Post-Training Detection of Backdoor Attacks with Arbitrary Backdoor Pattern Types Using a Maximum Margin Statistic
  • Citing Conference Paper
  • May 2024

... In [5], the study provides significant insights into resource management strategies in cloud computing, especially for dynamic and long-running workloads. It highlights the complexities of managing resources for applications such as query stream processing and distributed deep learning, emphasizing the need for efficient virtualized service selection and scaling of containerized microservices. ...

Online VM Service Selection with Spot Cores for Dynamic Workloads
  • Citing Conference Paper
  • June 2024

... These deep learning methods have significantly benefited our humans in daily life and created great economic outcomes in society. Along with their impressive development and great achievement, DNNs techniques have also exposed their untrustworthy aspects [13]- [17], which might perform unreliable predictions and cause severe economic, social, and security consequences, especially DNNs models, increasing efforts have been made to help users understand the inner working mechanism of DNNs models by providing interpretations on how certain decisions are made [27]- [30]. More specifically, the interpretations can be achieved by identifying the most influential parts (e.g., attribution maps) of the input with respect to its prediction [31]- [36]. ...

BIC-Based Mixture Model Defense Against Data Poisoning Attacks on Classifiers: A Comprehensive Study
  • Citing Article
  • August 2024

IEEE Transactions on Knowledge and Data Engineering

... It transforms regular triggers into noise triggers that can be easily hidden within images used to train neural networks, thereby implanting backdoors. Li et al. [119] introduced an easy but powerful backdoor attack targeting video data. The proposed attack adds perturbations in the transform domain, embedding imperceptible, temporally distributed triggers within video frames, and has been shown to be resilient against existing defense strategies. ...

Temporal-Distributed Backdoor Attack against Video Based Action Recognition
  • Citing Article
  • March 2024

Proceedings of the AAAI Conference on Artificial Intelligence

... On the other hand, truly unsupervised detection methods do not require labeled examples of what is normal and what is anomalous, and are analogous to unsupervised clustering methods. They model data distributions and flag potential outliers, e.g., [5,26,45,48] and our method proposed herein. The false positive rate (i.e., the fraction of normal samples misidentified as outliers) and false negative rate (i.e., the fraction of true outliers misidentified as normalities) may be relatively high for an unsupervised detector. ...

A BIC-Based Mixture Model Defense Against Data Poisoning Attacks on Classifiers
  • Citing Conference Paper
  • September 2023

... Scalability and Industrial Adoption: Although we consider model sizes which can fit within one GPU, as our proposed techniques optimize the embedding table granularity, our solutions are applicable for large-scale distributed inference scenarios [28]. Further, the forward pass in the training pipeline [81]- [83] could benefit from our schemes. By offering a readily deployable and performant solutions with prefetching and pinning, our work opens doors for wider industrial adoption of optimized DLRM inference pipelines. ...

Stash: A Comprehensive Stall-Centric Characterization of Public Cloud VMs for Distributed Deep Learning
  • Citing Conference Paper
  • Full-text available
  • July 2023

... Representation learning received limited attention in the context of backdoor defenses. Initial works design heuristics that leverage self-supervised representations to remove poisoned labels [33] or filter the dataset [85]. These heuristics fail for some attacks, as indicated by our experimental evaluation. ...

Training Set Cleansing of Backdoor Poisoning by Self-Supervised Representation Learning
  • Citing Conference Paper
  • June 2023

... By reversing the input and output of ordinary cGAN, the model can be used as a predictive model successfully. Wang, Miller & Kesidis (2023) proposed an unsupervised attack detector for DNN classifiers utilizing class-conditional GANs and modeling the distribution of clean data based on the predicted class label through an Auxiliary Classifier GAN (AC-GAN). Contreras-Cruz et al. (2023) developed the fast Anomaly Generative Adversarial Network (f-AnoGAN), a GAN architecture that exhibits superior accuracy when compared to bi-directional GAN and deep convolutional autoencoder. ...

Anomaly detection of adversarial examples using class-conditional generative adversarial networks
  • Citing Article
  • January 2023

Computers & Security

... where ρ = oi ϱi depicts the traffic intensity. Proof: The proof can be obtained by inverting the Laplace transform of S gap (t) at any time t = 0, given as [32], [33] E[e −sSgap(0) ] = ϱ i ϱ i + s ...

Age of information using Markov-renewal methods

Queueing Systems

... Simultaneously, the significant operational expenses of data centers have remained a challenge for major cloud service providers, leading to a continual focus on enhancing cost-effectiveness. In response, hybrid scheduling [2][3] [4] has emerged as a favored strategy, involving the simultaneous schedule of different types of workloads on the same node to improve resource utilization and reduce operational costs in cloud data centers.However, a schedule environment involves container-based workloads and the dynamics and diversity of different schedule environments, which brings complexity in resource scheduling and management. On the premise of ensuring the service quality of the workload, how to reduce the operating cost of the data center as much as possible through a reasonable scheduling mechanism has become a research problem. ...

Splice: An Automated Framework for Cost-and Performance-Aware Blending of Cloud Services
  • Citing Conference Paper
  • May 2022