Hatem Hajri’s research while affiliated with Innovation Research and Training Inc and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (55)


Proposed weather-robust cylinder counting pipeline
Illustration of PReNet architecture, a progressive network composed of a convolution layer fin\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$f_{in}$$\end{document} followed by an LSTM layer frecurrent\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$f_{recurrent}$$\end{document} and ResBloacks fres\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$f_{res}$$\end{document} into convolution layer output fout\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$f_{out}$$\end{document}. [26]
Illustration of ConvMixer architecture [29]
Illustration of examples of the original image and its corruption with different intensities of snowfall
Illustration of the PreNet purification through different stages on a clean image

+2

Real-time weather monitoring and desnowification through image purification
  • Article
  • Publisher preview available

April 2024

·

24 Reads

·

1 Citation

AI and Ethics

Eliott Py

·

Elies Gherbi

·

Nelson Fernandez Pinto

·

[...]

·

Hatem Hajri

Object detection and tracking are essential for reliable decision-making in modern applications, such as self-driving cars, drones, and industry. Adverse weather can hinder object detectability and pose a threat to the reliability of these systems. As a result, there is an increasing need for efficient image denoising and restoration techniques. In this study, we investigate the use of image purification as a means of defending against weather corruptions. Specifically, we focus on the effect of snow on an object detector and the benefits of efficient desnowification. We find that the performance of a strong image purifying baseline (PreNet) is not constant across different levels of snow intensity, leading to a reduced overall performance in diverse situations. Through extensive experimentation, we demonstrate that adding a lightweight snow detector significantly improves the overall object detection performance without needing to modify the purification model. Our proposed weather-robust architecture exhibits a 40% performance improvement compared to a strong image purification baseline on the gas cylinder counting task. In addition, it leads to significant reductions of the computational power required to run the purification pipeline with a minimal added cost.

View access options

SEEDS: Exponential SDE Solvers for Fast High-Quality Sampling from Diffusion Models

May 2023

·

266 Reads

A potent class of generative models known as Diffusion Probabilistic Models (DPMs) has become prominent. A forward diffusion process adds gradually noise to data, while a model learns to gradually denoise. Sampling from pre-trained DPMs is obtained by solving differential equations (DE) defined by the learnt model, a process which has shown to be prohibitively slow. Numerous efforts on speeding-up this process have consisted on crafting powerful ODE solvers. Despite being quick, such solvers do not usually reach the optimal quality achieved by available slow SDE solvers. Our goal is to propose SDE solvers that reach optimal quality without requiring several hundreds or thousands of NFEs to achieve that goal. In this work, we propose Stochastic Exponential Derivative-free Solvers (SEEDS), improving and generalizing Exponential Integrator approaches to the stochastic case on several frameworks. After carefully analyzing the formulation of exact solutions of diffusion SDEs, we craft SEEDS to analytically compute the linear part of such solutions. Inspired by the Exponential Time-Differencing method, SEEDS uses a novel treatment of the stochastic components of solutions, enabling the analytical computation of their variance, and contains high-order terms allowing to reach optimal quality sampling 3\sim3-5×5\times faster than previous SDE methods. We validate our approach on several image generation benchmarks, showing that SEEDS outperforms or is competitive with previous SDE solvers. Contrary to the latter, SEEDS are derivative and training free, and we fully prove strong convergence guarantees for them.



A hyperbolic approach for learning communities on graphs

February 2023

·

260 Reads

·

3 Citations

Data Mining and Knowledge Discovery

Detecting communities on graphs has received significant interest in recent literature. Current state-of-the-art approaches tackle this problem by coupling Euclidean graph embedding with community detection. Considering the success of hyperbolic representations of graph-structured data in the last years, an ongoing challenge is to set up a hyperbolic approach to the community detection problem. The present paper meets this challenge by introducing a Riemannian geometry based framework for learning communities on graphs. The proposed methodology combines graph embedding on hyperbolic spaces with Riemannian K-means or Riemannian mixture models to perform community detection. The usefulness of this framework is illustrated through several experiments on generated community graphs and real-world social networks as well as comparisons with the most powerful baselines. The code implementing hyperbolic community embedding is available online https://www.github.com/tgeral68/HyperbolicGraphAndGMM.


Neural Adversarial Attacks with Random Noises

December 2022

·

16 Reads

·

1 Citation

International Journal of Artificial Intelligence Tools

In this paper, we present an approach which relies on the use of random noises to generate adversarial examples of deep neural network classifiers. We argue that existing deterministic attacks, which perform by sequentially applying maximal perturbations on selected components of the input, fail at reaching accurate adversarial examples on real-world large scale datasets. By exploiting a simple Taylor expansion of the expected output probability under the noise perturbation, we introduce noise-based sparse (or L 0 ) targeted and untargeted attacks. Our proposed method, called Voting Folded Gaussian Attack (VFGA), achieves significantly better L 0 scores than state-of-the-art L 0 attacks (such as SparseFool and Sparse-RS) while being faster on both CIFAR-10 and ImageNet. Moreover, we show that VFGA is also applicable as an L ∞ attack and outperforms the state-of-the-art projected gradient attack (PGD) method.


Improving Robustness of Deep Reinforcement Learning Agents: Environment Attack based on the Critic Network

July 2022

·

62 Reads

·

4 Citations

To improve policy robustness of deep reinforcement learning agents, a line of recent works focus on producing disturbances of the environment. Existing approaches of the literature to generate meaningful disturbances of the environment are adversarial reinforcement learning methods. These methods set the problem as a two-player game between the protagonist agent, which learns to perform a task in an environment, and the adversary agent, which learns to disturb the protagonist via modifications of the considered environment. Both protagonist and adversary are trained with deep reinforcement learning algorithms. Alternatively, we propose in this paper to build on gradient-based adversarial attacks, usually used for classification tasks for instance, that we apply on the critic network of the protagonist to identify efficient disturbances of the environment. Rather than learning an attacker policy, which usually reveals as very complex and unstable, we leverage the knowledge of the critic network of the protagonist, to dynamically complexify the task at each step of the learning process. We show that our method, while being faster and lighter, leads to significantly better improvements in policy robustness than existing methods of the literature.


Riemannian data-dependent randomized smoothing for neural networks certification

June 2022

·

65 Reads

Certification of neural networks is an important and challenging problem that has been attracting the attention of the machine learning community since few years. In this paper, we focus on ran-domized smoothing (RS) which is considered as the state-of-the-art method to obtain certifiably robust neural networks. In particular, a new data-dependent RS technique called ANCER introduced recently can be used to certify ellipses with orthogonal axis near each input data of the neural network. In this work, we remark that ANCER is not invariant under rotation of input data and propose a new rotationally-invariant formulation of it which can certify ellipses without constraints on their axis. Our approach called Riemannian Data Dependant Randomized Smoothing (RD-DRS) relies on information geometry techniques on the manifold of covariance matrices and can certify bigger regions than ANCER based on our experiments on the MNIST dataset.


Mean A rel c (%) at changes in corruption severity for MNIST. At fixed (c, s) ∈ C, each block (in green) contains results for cleanly trained ResNet (upper-left), ODENet (lower-left) and their noisy counterparts (upper-right, lower-right) The listed corruptions are as in Table 1. Performance shifts are colored in red. Corruptions where noisy training is not beneficial are colored in blue.
Mean A rel c (%) on corrupted SVHN images for ResNet and ODENet. The listed corruptions are as in Table 1. The last block computes the improvement on performance for each model induced by noisy training w.r.t. clean training. Corruptions where noisy training is not beneficial are colored in blue and a > 5% difference of model's performance is colored in orange.
Mean A rel c (%) at changes in severity. The listed corruptions are as in Table 1. Red color means a shift of best model's accuracy w.r.t. previous severity value. Corruptions where noisy training is not beneficial are colored in blue.
Noisy Learning for Neural ODEs Acts as a Robustness Locus Widening

June 2022

·

56 Reads

We investigate the problems and challenges of evaluating the robustness of Differential Equation-based (DE) networks against synthetic distribution shifts. We propose a novel and simple accuracy metric which can be used to evaluate intrinsic robustness and to validate dataset corruption simulators. We also propose methodology recommendations, destined for evaluating the many faces of neural DEs' robustness and for comparing them with their discrete counterparts rigorously. We then use this criteria to evaluate a cheap data augmentation technique as a reliable way for demonstrating the natural robustness of neural ODEs against simulated image corruptions across multiple datasets.


Realization Theory Of Recurrent Neural ODEs Using Polynomial System Embeddings

May 2022

·

35 Reads

In this paper we show that neural ODE analogs of recurrent (ODE-RNN) and Long Short-Term Memory (ODE-LSTM) networks can be algorithmically embeddeded into the class of polynomial systems. This embedding preserves input-output behavior and can suitably be extended to other neural DE architectures. We then use realization theory of polynomial systems to provide necessary conditions for an input-output map to be realizable by an ODE-LSTM and sufficient conditions for minimality of such systems. These results represent the first steps towards realization theory of recurrent neural ODE architectures, which is is expected be useful for model reduction and learning algorithm analysis of recurrent neural ODEs.



Citations (26)


... (4) Lighting and weather conditions: Real-world data is complex and variable because it includes good weather conditions as well as bad weather (e.g., rainy, snowy etc.) [210] and bad lighting conditions (e.g., foggy, dark, night etc.) [211]. These adverse conditions affect the appearance of objects in the scene, negatively affecting the accuracy of semantic segmentation. ...

Reference:

Deep-Learning-Based Approaches for Semantic Segmentation of Natural Scene Images: A Review
Real-time weather monitoring and desnowification through image purification

AI and Ethics

... Future work should involve the testing of clustering methods based on near-neighbour graphs and more complex MST-inspired data structures González-Barrios & Quiroz, 2003;Karypis et al., 1999;Zhong et al., 2011Zhong et al., , 2010. Also, the application of the MST-based algorithms could be examined in the problem of community detection in graphs (Gerald et al., 2023). ...

A hyperbolic approach for learning communities on graphs

Data Mining and Knowledge Discovery

... While several innovative adversarial attack methods [11,28,2], and efficient defense mechanisms [18,1,7] exist, we demonstrate how just changing the training algorithm can affect the robustness of a simple adversarial attack. Our contributions in summary • We show, experimentally, how training a two-layer neural network with adaptive random Fourier features [13] can overcome the spectral bias, as defined in [16], compared to training with stochastic gradient descent. ...

Neural Adversarial Attacks with Random Noises
  • Citing Article
  • December 2022

International Journal of Artificial Intelligence Tools

... As done in works crafting attacks on observations given an actor policy [9], we seek at crafting an adversarial observation x ′ , by applying a perturbation η on x, that minimizes the probability of the dominant action a d . The principle of our method, inspired by [16] (see also [4] for a recent improvement), is to leverage the knowledge of the Jacobian matrix of the function learned by the actor network with respect to the inputs to generate perturbations. Consider the actor network Π θ , and denote by Π(x) its logit outputs, Π aj (x) is the probability of the action a j given the input x. ...

Stochastic sparse adversarial attacks
  • Citing Conference Paper
  • November 2021

... Thus, it could be said that the two can complement each other to form a fairly robust obstacle detection method. In this way, there are many works in which multiple sensors are used for obstacle detection [3][4][5][6]. ...

Real Time Lidar and Radar High-Level Fusion for Obstacle Detection and Tracking with Evaluation on a Ground Truth
  • Citing Article
  • July 2018

International Journal of Mechanical & Mechatronics Engineering

... Enhanced Software and Methodological Tools. Several libraries currently exist for statistics and optimization on Riemannian manifolds, such as Geomstats [133] and Geoopt [107], as well as libraries for distance metric learning, including pydml [168] and PyTorch Metric Learning [137], particularly in the context of deep learning. However, to date, no comprehensive library specifically tailored for Riemannian metric learning has been developed. ...

Geomstats: A Python Package for Riemannian Geometry in Machine Learning

... The one-pixel attack achieves good attack performance on the adversarial samples generated by perturbing only one pixel on the Kaggle CIFAR-10 dataset. Papernot et al. [21,22] proposed an adversarial attack method called JSMA, which perturbs image samples with the L 0 norm distance as a constraint. This method builds a saliency map based on the forward derivative to reflect which pixels have a greater impact on the image. ...

Probabilistic Jacobian-Based Saliency Maps Attacks

Machine Learning and Knowledge Extraction

... We emphasize that we have a different equation for every u ∈ R, but all depend on the same process W = (X, Y). For a fixed u ∈ R, this equation, that goes under the name of perturbed Tanaka's SDE, has already been studied in the literature [Pro13,cHK18] in the case where W is a two-dimensional Brownian motion of correlation ρ with ρ ∈ (−1, 1), and more generally when the correlation coefficient varies with time. In particular, pathwise uniqueness and existence of a strong solution are known. ...

Correlated Coalescing Brownian Flows on R and the Circle
  • Citing Article
  • January 2018

Latin American Journal of Probability and Mathematical Statistics