
Georgios AnagnostopoulosFlorida Institute of Technology · Department of Electrical & Computer Engineering
Georgios Anagnostopoulos
Ph.D. in Electrical Engineering
About
144
Publications
18,033
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
1,554
Citations
Citations since 2017
Introduction
My research interests lie in the theory and applications of Machine Learning.
Additional affiliations
May 2012 - June 2012
October 2011 - December 2011
August 2008 - present
Education
June 1997 - August 2001
May 1995 - May 1997
September 1987 - May 1994
Publications
Publications (144)
One important aspect of understanding behaviors of information cascades is to be able to accurately predict their popularity, that is, their message counts at any future time. Self-exciting Hawkes processes have been widely adopted for such tasks due to their success in describing cascading behaviors. In this paper, for general, marked Hawkes point...
Flood prediction across scales and more specifically in ungauged areas remains a great challenge that limits the efficiency of flood risk mitigation strategies and disaster preparedness. Building upon the recent success of Machine Learning (ML) models on streamflow prediction, this work presents a prototype ML-based framework for flood warning and...
It is widely understood that diffusion of and simultaneous interactions between narratives -- defined here as persistent point-of-view messaging -- significantly contributes to the shaping of political discourse and public opinion. In this work, we propose a methodology based on Multi-Variate Hawkes Processes and our newly-introduced Process Influe...
This paper explains the design of a social network analysis framework, developed under DARPA’s SocialSim program, with novel architecture that models human emotional, cognitive, and social factors. Our framework is both theory and data-driven, and utilizes domain expertise. Our simulation effort helps understanding how information flows and evolves...
It is widely understood that diffusion of and simultaneous interactions between narratives—defined here as persistent point-of-view messaging—significantly contributes to the shaping of political discourse and public opinion. In this work, we propose a methodology based on Multi-Variate Hawkes Processes and our newly-introduced Process Influence Me...
This paper explains the design of a social network analysis framework, developed under DARPA's SocialSim program, with novel architecture that models human emotional, cognitive and social factors. Our framework is both theory and data-driven, and utilizes domain expertise. Our simulation effort helps in understanding how information flows and evolv...
In this paper, we propose a novel hash learning approach that has the following main distinguishing features, when compared to past frameworks. First, the codewords are utilized in the Hamming space as ancillary techniques to accomplish its hash learning task. These codewords, which are inferred from the data, attempt to capture grouping aspects of...
We propose a new method for local distance metric learning based on sample similarity as side information. These local metrics, which utilize conical combinations of metric weight matrices, are learned from the pooled spatial characteristics of the data, as well as the similarity profiles between the pairs of samples, whose distances are measured....
We show a Talagrand-type concentration inequality for Multi-Task Learning (MTL), with which we establish sharp excess risk bounds for MTL in terms of the Local Rademacher Complexity (LRC). We also give a new bound on the LRC for any norm regularized hypothesis classes, which applies not only to MTL, but also to the standard Single-Task Learning (ST...
This paper introduces a new and effective algorithm for learning kernels in a Multi-Task Learning (MTL) setting. Although, we consider a MTL scenario here, our approach can be easily applied to standard single task learning, as well. As shown by our empirical results, our algorithm consistently outperforms the traditional kernel learning algorithms...
In machine learning, the notion of multi-objective model selection (MOMS) refers to the problem of identifying the set of Pareto-optimal models that optimize by compromising more than one predefined objectives simultaneously. This paper introduces SPRINT-Race, the first multi-objective racing algorithm in a fixed-confidence setting, which is based...
In this paper we introduce a novel hash learning framework that has two main distinguishing features, when compared to past approaches. First, it utilizes codewords in the Hamming space as ancillary means to accomplish its hash learning task. These codewords, which are inferred from the data, attempt to capture similarity aspects of the data's hash...
When faced with learning a set of inter-related tasks from a limited amount
of usable data, learning each task independently may lead to poor
generalization performance. Multi-Task Learning (MTL) exploits the latent
relations between tasks and overcomes data scarcity limitations by co-learning
all these tasks simultaneously to offer improved perfor...
In this paper we introduce a novel hash learning framework that has two main distinguishing features, when compared to past approaches. First, it utilizes codewords in the Hamming space as ancillary means to accomplish its hash learning task. These codewords, which are inferred from the data, attempt to capture similarity aspects of the data's hash...
Model selection is a core aspect in machine learning and is, occasionally, multi-objective in nature. For instance, hyper-parameter selection in a multi-task learning context is of multi-objective nature, since all the tasks' objectives must be optimized simultaneously. In this paper, a novel multi-objective racing algorithm (RA), namely S-Race, is...
In this paper we introduce a novel hash learning framework that has two main distinguishing features, when compared to past approaches. First, it utilizes codewords in the Hamming space as ancillary means to accomplish its hash learning task. These codewords, which are inferred from the data, attempt to capture similarity aspects of the data's hash...
Multi-objective model selection, which is an important aspect of Machine Learning, refers to the problem of identifying a set of Pareto optimal models from a given ensemble of models. This paper proposes SPRINT-Race, a multi-objective racing algorithm based on the Sequential Probability Ratio Test with an Indifference Zone. In SPRINT-Race, a non-pa...
The support vector machine (SVM) remains a popular classifier for its excellent generalization performance and applicability of kernel methods; however, it still requires tuning of a regularization parameter, C, to achieve optimal performance. Regularization path-following algorithms efficiently solve the solution at all possible values of the regu...
A traditional and intuitively appealing Multi-Task Multiple Kernel Learning
(MT-MKL) method is to optimize the sum (thus, the average) of objective
functions with (partially) shared kernel function, which allows information
sharing amongst tasks. We point out that the obtained solution corresponds to a
single point on the Pareto Front (PF) of a Mul...
Traditionally, Multi-task Learning (MTL) models optimize the average of
task-related objective functions, which is an intuitive approach and which we
will be referring to as Average MTL. However, a more general framework,
referred to as Conic MTL, can be formulated by considering conic combinations
of the objective functions instead; in this framew...
Racing algorithms are often used for offline model selection, where models are compared in terms of their average performance over a collection of problems. In this paper, we present a new racing algorithm variant, Max-Race, which makes decisions based on the maximum performance of models. It is an online algorithm, whose goal is to optimally alloc...
In this paper we introduce a derivative-free optimization method that is derived from a population based stochastic gradient estimator. We first demonstrate some properties of this estimator and show how it is expected to always yield a descent direction. We analytically show that the difference between the expected function value and the optimum d...
Over the past few years, Multi-Kernel Learning (MKL) has received significant
attention among data-driven feature selection techniques in the context of
kernel-based learning. MKL formulations have been devised and solved for a
broad spectrum of machine learning problems, including Multi-Task Learning
(MTL). Solving different MKL formulations usual...
This paper presents a RKHS, in general, of vector-valued functions intended
to be used as hypothesis space for multi-task classification. It extends
similar hypothesis spaces that have previously considered in the literature.
Assuming this space, an improved Empirical Rademacher Complexity-based
generalization bound is derived. The analysis is itse...
We propose a new method for local metric learning based on a conical combination of Mahalanobis metrics and pair-wise similar-ities between the data. Its formulation allows for controlling the rank of the metrics' weight matrices. We also offer a convergent algorithm for training the associated model. Experimental results on a collection of classif...
In this paper we present two related, kernelbased Distance Metric Learning (DML) methods. Their respective models non-linearly map data from their original space to an output space, and subsequent distance measurements are performed in the output space via a Mahalanobis metric. The dimensionality of the output space can be directly controlled to fa...
This paper presents a multi-objective racing algorithm, S-Race, which efficiently addresses multi-objective model selection problems in the sense of Pareto optimality. As a racing algorithm, S-Race attempts to eliminate candidate models as soon as there is sufficient statistical evidence of their inferiority relative to other models with respect to...
Existing active set methods reported in the literature for support vector machine (SVM) training must contend with singularities when solving for the search direction. When a singularity is encountered, an infinite descent direction can be carefully chosen that avoids cycling and allows the algorithm to converge. However, the algorithm implementati...
Over the last few years, Kernel Principal Component Analysis (KPCA) has found several applications in outlier detection. A relatively recent method uses KPCA to compute the reconstruction error (RE) of previously unseen samples and, via thresholding, to identify atypical samples. In this paper we propose an alternative method, which performs the sa...
The Sammon Mapping (SM) has established itself as a valuable tool in dimensionality reduction, manifold learning, exploratory data analysis and, particularly, in data visualization. The SM is capable of projecting high-dimensional data into a low-dimensional space, so that they can be visualized and interpreted. This is accomplished by representing...
In this paper we introduce Multinomial Squared Direction Cosines Regression as an alternative Multinomial Response Model. The proposed model offers an intuitive geometric interpretation to the task of estimating posterior class probabilities in multi-class problems. In specific, the latter probabilities correspond to the squared direction cosines b...
In this paper we present a novel framework for evolving ART-based classification models, which we refer to as MOME-ART. The new training framework aims to evolve populations of ART classifiers to optimize both their classification error and their structural complexity. Towards this end, it combines the use of interacting sub-populations, some tradi...
In this paper we present a novel generalization of Sammon's Mapping (SM), which is a popular, metric multi-dimensional scaling technique used in data analysis and visualization. The new approach, namely the Kernel-based Sammon Mapping (KSM), yields the classic SM and other much related techniques as special cases. Apart from being able to approxima...
In this paper, we present the evolution of adaptive resonance theory (ART) neural network architectures (classifiers) using a multiobjective optimization approach. In particular, we propose the use of a multiobjective evolutionary approach to simultaneously evolve the weights and the topology of three well-known ART architectures; fuzzy ARTMAP (FAM...
The work in this paper explores the discriminatory power of target outline description features in conjunction with Support Vector Machine (SVM) based classification committees, when attempting to recognize a variety of targets from Synthetic Aperture Radar (SAR) images. In specific, approximate target outlines are first determined from SAR images...
Real-time, static and dynamic hand gesture learning and recognition makes it possible to have computers recognize hand gestures naturally. This creates endless possibilities in the way humans can interact with computers, allowing a human hand to be a peripheral by itself. The software framework developed provides a lightweight, robust, and practica...
Efficiently implemented active set methods have been successfully applied to Support Vector Machine (SVM) training. These active set methods offer higher precision and incremental training at the cost of additional memory requirements when compared to decomposition methods such as Sequential Minimal Optimization (SMO). However, all existing active...
This paper focuses on classification problems, and in particular on the evolution of ARTMAP architectures using genetic algorithms, with the objective of improving generalization performance and alleviating the adaptive resonance theory (ART) category proliferation problem. In a previous effort, we introduced evolutionary fuzzy ARTMAP (FAM), referr...
As the realism in games continues to increase, through improvements in graphics and 3D engines, more focus is placed on the behavior of the simulated agents that inhabit the simulated worlds. The agents in modern video games must become more life-like in order to seem to belong in the environments they are portrayed in. Many modern artificial intel...
Active set methods for training the support vector machines (SVM) are advantageous since they enable incremental training and, as we show in this research, do not exhibit exponentially increasing training times commonly associated with the decomposition methods as the SVM training parameter, C, is increased or the classification difficulty increase...
The AMALTHEA REU Program is a 10-week, summer research experience for science or engineering undergraduate students funded by the National Science Foundation since 2007 and featuring Machine Learning as its intellectual focus. Moreover, it is a joint effort of two collaborating universities in Central Florida, namely Florida Institute of Technology...
Pruning is one of the key procedures in training de- cision tree classifiers. It removes trivial rules from the raw knowledge base built from training examples, in order to avoid over-using noisy, conflicting, or fuzzy inputs, so that the refined model can generalize better with unseen cases. In this paper, we present a num- ber of properties of k-...
The joint workshops aim at promoting interaction and collaboration not only among researchers working directly in areas covered by TC1 and TC2 but also among those in other fields who use statistical, structural or syntactic techniques extensively. We welcome mathematicians, statisticians, researchers in machine learning and practitioners alike who...
The joint workshops aim at promoting interaction and collaboration not only among researchers working directly in areas covered by TC1 and TC2 but also among those in other fields who use statistical, structural or syntactic techniques extensively. We welcome mathematicians, statisticians, researchers in machine learning and practitioners alike who...
In this work we present, for the first time, the evolution of ART Neural Network architectures (classifiers) using a multiobjective optimization approach. In particular, we propose the use of a multiobjective evolutionary approach to evolve simultaneously the weights, as well as the topology of three well-known ART architectures; Fuzzy ARTMAP (FAM)...
Genetic algorithms have been used to evolve several neural network architectures. In a previous effort, we introduced the evolution of three well known ART architects; Fuzzy ARTMAP (FAM), Ellipsoidal ARTMAP (EAM) and Gaussian ARTMAP (GAM). The resulting architectures were shown to achieve competitive generalization and exceptionally small size. A m...
This paper primarily investigates the use of shape-based features by an Automatic Target Recognition (ATR) system to classify various types of targets in Synthetic Aperture Radar (SAR) images. In specific, shapes of target outlines are represented via Elliptical Fourier Descriptors (EFDs), which, in turn, are utilized as recognition features. Accor...
Decision trees are well-known and established models for classification and regression. In this paper, we focus on the estimation
and the minimization of the misclassification rate of decision tree classifiers. We apply Lidstone’s Law of Succession for
the estimation of the class probabilities and error rates. In our work, we take into account not...
As massively multi-player gaming environments become more detailed, developing agents to populate these virtual worlds as capable non-player characters poses an increas- ingly complex problem. Human players in many games must achieve their objectives through financial skills such as tr ad- ing and supply chain management as well as through com- bat...
In machine learning, decision trees are employed extensively in solving classification problems. In order to design a decision tree classifier two main phases are employed. The first phase is to grow the tree using a set of data, called training data, quite often to its maximum size. The second phase is to prune the tree. The pruning phase produces...
Several sets of features, existent in triangulated, irregularly spaced LiDAR data, are extracted, conditioned, and presented to a number of clustering algorithms with the intent to recognize planar structures within the data. From those planar structures, encoded by the clustering algorithms, 3D models are then reconstructed. The purpose of this pa...
This book constitutes the refereed proceedings of the 12th International Workshop on Structural and Syntactic Pattern Recognition, SSPR 2008 and the 7th International Workshop on Statistical Techniques in Pattern Recognition, SPR 2008, held jointly in Orlando, FL, USA, in December 2008 as a satellite event of the 19th International Conference of Pa...
Outlier detection has received significant attention in many applications, such as detecting credit card fraud or network intrusions. Most existing research focuses on numerical datasets, and cannot directly apply to categorical sets where there is little sense in calculating distances among data points. Furthermore, a number of outlier detection m...
This paper focuses on the evolution of Fuzzy ARTMAP neural network classifiers, using genetic algorithms, with the objective of improving generalization performance (classification accuracy of the ART network on unseen test data) and alleviating the ART category proliferation problem (the problem of creating more than necessary ART network categori...
Probabilistic neural networks (PNN) and general regression neural networks (GRNN) represent knowledge by simple but interpretable models that approximate the optimal classifier or predictor in the sense of expected value of the accuracy. These models require the specification of an important smoothing parameter, which is usually chosen by cross-val...
The support vector machine is a widely employed machine learning model due to its repeatedly demonstrated superior generalization performance. The sequential minimal optimization (SMO) algorithm is one of the most popular SVM training approaches. SMO is fast, as well as easy to implement; however, it has a limited working set size (2 points only)....
This paper focuses on the evolution of ARTMAP architectures, using genetic algorithms, with the objective of improving generalization performance and alleviating the ART category proliferation problem. We refer to the resulting architectures as GFAM, GEAM, and GGAM. We demonstrate through extensive experimentation that evolved ARTMAP architectures...
Adaptive Resonance Theory (ART) neural network architectures, such as Fuzzy ARTMAP (FAM), have solved successfully a variety of classification problems. However, FAM suffers from an inherent problem that of creating larger architectures than it is necessary to solve the problem at hand (referred to as the ART category proliferation problem). This p...
Introduction Methods and Systems Experimental Results Conclusions References