Marco Lippi’s research while affiliated with University of Florence and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (104)


Figure 1: Task examples. Left: positive (a) and negative (b) examples for task 12 in KANDY-1, positive class being defined by a red triangle on the right, plus the presence of a blue object anywhere. Right: positive (c) and negative (d) examples for task 24 in KANDY-2, positive class being defined by objects that share one attribute (in this case, color).
Correlations between discovered concepts and (expected) ground truth concepts. The setting with the largest diagonal score (DIAGS) and at least another winning metric is highlighted in gray. Results are shown as mean ± std across 5 runs.
Continual Learning For Unsupervised Concept Bottleneck Discovery
  • Conference Paper
  • Full-text available

July 2024

Luca Salvatore Lorello

·

Marco Lippi

·

In the context of continual learning, little attention is dedicated to the problem of developing a layer of "concepts", also known as "concept bottleneck", to support the discrimination of higher-level task information, especially when concepts are not supervised. Concept bottleneck discovery in an unsupervised setting is thus largely unexplored, and this paper aims to move a step forward in such direction. We consider a neural network that faces a stream of binary tasks, with no further information on the relationships among them, i.e., no supervisions at the level of concepts. The learning of the concept bottleneck layer is driven by means of a triplet-based criterion, which is instantiated in conjunction with a specifically designed experience replay (concept replay). Such a novel criterion exploits fuzzy Hamming distances to treat vectors of concept probabilities as fuzzy bitstrings, encouraging different concept activations across different tasks, while also adding a regularization effect which pushes probabilities towards crisp values. Despite the lack of concept supervisions, we found that continually learning the streamed tasks in a progressive manner yields the development of inner concepts that are significantly better correlated with the higher-level tasks, compared to the case of joint-offline learning. This result is showcased in an extended experimental activity involving different architectures and newly created (and shared) datasets that are also well-suited to support further investigation of continual learning in concept-based models.

Download

The four alternatives tested in our approach, exemplified for the German language. From the top left corner, clockwise: (1) re-train from scratch a German version of CLAUDETTE, with an original German corpus; (2) project labels from English to German documents, and re-train a German CLAUDETTE; (3) translate documents from English to German, keep the original English annotations, and re-train a German CLAUDETTE; (4) use the English CLAUDETTE, and translate query documents from German
Unfair clause detection in terms of service across multiple languages

April 2024

·

39 Reads

·

1 Citation

Artificial Intelligence and Law

Most of the existing natural language processing systems for legal texts are developed for the English language. Nevertheless, there are several application domains where multiple versions of the same documents are provided in different languages, especially inside the European Union. One notable example is given by Terms of Service (ToS). In this paper, we compare different approaches to the task of detecting potential unfair clauses in ToS across multiple languages. In particular, after developing an annotated corpus and a machine learning classifier for English, we consider and compare several strategies to extend the system to other languages: building a novel corpus and training a novel machine learning system for each language, from scratch; projecting annotations across documents in different languages, to avoid the creation of novel corpora; translating training documents while keeping the original annotations; translating queries at prediction time and relying on the English system only. An extended experimental evaluation conducted on a large, original dataset indicates that the time-consuming task of re-building a novel annotated corpus for each language can often be avoided with no significant degradation in terms of performance.



Explaining population variation after the 2016 Central Italy earthquake using Call Data Records and Twitter

October 2023

·

26 Reads

·

1 Citation

Social Network Analysis and Mining

In this work, we analyse mobile phone variation before and after the 2016 Central Italy earthquake in the affected areas, using Twitter and public reconstruction works data. We create three models and show that Twitter data and the related sentiment on the earthquake, as well as the distribution of emergency houses, can contribute to explaining population variations. Our final Generalised Poisson regression model explains more than 80% of the variance of the population’s variation based on the percentage of negative polarity tweets, the number of emergency houses, the number of negative tweets on the earthquake weighted by the number of residents, number of tweets posted on the earthquake anniversary, the distance from the epicentre and several variables related to public reconstruction works (e.g. school, public housing, hydrological disruption, viability). We found that sentiment on the emergency house can be a proxy for population variation because people who live there did not displace from the crater area. The number of tweets posted during the anniversary day can, instead, indicate negative population variation because the higher the number of tweets, the more people can feel nostalgic after having relocated.


Setup Time Prediction Using Machine Learning Algorithms: A Real-World Case Study

September 2023

·

102 Reads

·

1 Citation

IFIP Advances in Information and Communication Technology

In this paper, we explore the use of machine learning regression algorithms for setup time prediction and we apply them to a real-world scheduling application arising in the color printing industry. As the complexities associated with setup time predictions have received limited attention from the literature, we aim at exploiting a data-driven approach based on machine learning algorithms to enhance the quality of setup time evaluations and narrow the gap between scheduling theory and practice. Using a real-world industrial dataset, we train three different machine learning models: linear regression, random forests, and gradient boosting machines. The experimental results demonstrate that the gradient boosting machine approach obtains the best performance overall, immediately followed by random forests. The accuracy of the obtained predictions shows the effectiveness of the proposed approach in setup time evaluation. The obtained results are particularly significant and valuable due to the versatile nature of the proposed machine learning approaches. These methods can be applied to various scheduling scenarios, making them suitable for integration into scheduling algorithms to potentially improve their accuracy.



Multi-Task Attentive Residual Networks for Argument Mining

May 2023

·

46 Reads

·

16 Citations

IEEE/ACM Transactions on Audio Speech and Language Processing

We explore the use of residual networks and neural attention for multiple argument mining tasks. We propose a residual architecture that exploits attention, multi-task learning, and makes use of ensemble, without any assumption on document or argument structure. We present an extensive experimental evaluation on five different corpora of user-generated comments, scientific publications, and persuasive essays. Our results show that our approach is a strong competitor against state-of-the-art architectures with a higher computational footprint or corpus-specific design, representing an interesting compromise between generality, performance accuracy and reduced model size.


A General Pipeline for Online Gesture Recognition in Human–Robot Interaction

April 2023

·

73 Reads

·

8 Citations

IEEE Transactions on Human-Machine Systems

Recent advances in robotics have allowed the introduction of robots assisting and working together with human subjects. To promote their use and diffusion, intuitive and user-friendly interaction means should be adopted. In particular, gestures have become an established way to interact with robots since they allow to command them in an intuitive manner. In this article, we focus on the problem of gesture recognition in human–robot interaction (HRI). While this problem has been largely studied in the literature, it poses specific constraints when applied to HRI. We propose a framework consisting in a pipeline devised to take into account these specific constraints. We implement the proposed pipeline considering, as an example, an evaluation use case. To this end, we consider standard machine learning algorithms for the classification stage and evaluate their performance considering different performance metrics for a thorough assessment.


Explaining population variation after the 2016 Central Italy earthquake using Call Data Records and Twitter

February 2023

·

23 Reads

In this work, we analyse mobile phone variation before and after the 2016 Central Italy earthquake in the affected areas, using Twitter and public reconstruction works data. We create three models and show that Twitter data and the related sentiment on the earthquake, as well as the distribution emergency houses, can contribute to explaining population variations. Our final Poisson regression model explains more than 80% of the variance of the population's variation based on the percentage of negative polarity tweets, the number of emergency houses, the number of negative tweets on the earthquake weighted by the number of residents, number of tweets posted on the earthquake anniversary, the distance from the epicentre and several variables related to public reconstruction works (e.g., school, public housing, hydrological disruption, viability). We found that sentiment on the emergency house can be a proxy for population variation because people who live there did not displace from the crater area. The number of tweets posted during the anniversary day can, instead, indicate negative population variation because the higher the number of tweets, the more people can feel nostalgic after having relocated.


FIGURE 1. Simplified drawing of bearing structure showing the characteristic dimensions.
Specifications of the 6205 ball bearing used in the experiments and expected fault frequencies
Accuracy for anomaly detection, 2 classes (pre-processing only).
Accuracy for fault recognition, 3 classes (pre-processing only).
Bearing Fault Detection and Recognition From Supply Currents With Decision Trees

January 2023

·

85 Reads

·

5 Citations

IEEE Access

This paper considers the tasks of detecting and recognizing bearing faults in electric motors from the signals collected from supply currents, using machine learning techniques. In particular, following recent trends in AI, the main point of interest was focused towards interpretable solutions that provide explanations on the decisions taken by the classifiers. For this reason, decision trees were chosen, since they represent a classic machine learning approach which inductively learns tree structures from a collection of observations. Paths along the learnt trees can be easily interpreted as plain classification rules. An extensive experimental comparison shows the strong generalization capabilities of such a classifier. In particular, the present work reports results obtained in a highly challenging scenario, usually overlooked in the literature, where the system is tested on configurations of radial and torsional loads that have not been observed during training. The proposed approach achieves over 90% of accuracy even on this cross-load generalization setting.


Citations (72)


... This method operates by considering the degree of similarity among nearest neighbors [31] Decision trees (DTs) DT is a learning algorithm frequently used in the literature for classification problems. The decision tree consists of branches, leaves and decision nodes [32] Support vector machine (SVM) SVM is a supervised algorithm utilized for classification or regression tasks. Primarily employed in classification problems, SVM operates based on the concept of a hyperplane that effectively separates different data classes [33] Convolutional neural network (CNN) CNNs, extensively employed in various computer vision domains, particularly in classification tasks, differ from classical neural networks by incorporating convolutional feature extraction and classification layers [34] 3 Experimental studies and results ...

Reference:

A novel hybrid deep learning model for early stage diabetes risk prediction
Bearing Fault Detection and Recognition From Supply Currents With Decision Trees

IEEE Access

... Twitter data has also been utilized for marketing and social studies, providing insights into people's opinions on political, religious, and societal events (Zhou et al. 2013;Mandloi and Patel 2020). A recent study (Hadjidimitriou et al. 2023) examined changes in mobile phone usage before and after the 2016 Central Italy earthquake in affected regions, utilizing Twitter activity and data on public reconstruction efforts. Three models are developed, illustrating how Twitter content and sentiments regarding the earthquake, alongside the distribution of emergency housing, helped elucidate fluctuations in population. ...

Explaining population variation after the 2016 Central Italy earthquake using Call Data Records and Twitter

Social Network Analysis and Mining

... Constraints (19) guarantee that, during the processing of job j ∈ J, each tool required by j takes exactly a slot. Constraints (20) guarantee that the precedence relation ≺ j at each job j is fulfilled. ...

Setup Time Prediction Using Machine Learning Algorithms: A Real-World Case Study

IFIP Advances in Information and Communication Technology

... Comprehensive surveys on DTs for IoT (Minerva et al. 2020) and smart manufacturing (Friederich et al. 2022;Lippi et al. 2023) have appeared in the literature that point to the ability to solve problems of a realistic size in a data-driven setting. Notably, Kritzinger et al. (2018) distinguish between a DT that considers machine-level impacts and a digital shadow that considers system-level impacts. ...

Enabling causality learning in smart factories with hierarchical digital twins
  • Citing Article
  • June 2023

Computers in Industry

... Some prior work also studies ArgMining across different corpora. Given the variability of annotation schemas, dealing with different conceptualizations (such as tree vs. graph-based structures, ADU and relation labels, ADU boundaries, among others) is a common challenge [2,10,22]. Besides the variability of annotated resources, ArgMining corpora tend to be small [41]. ...

Multi-Task Attentive Residual Networks for Argument Mining

IEEE/ACM Transactions on Audio Speech and Language Processing

... Combining legal and AI tools offers considerable untapped potential in the advancement of consumer protection . Although several technical solutions have emerged (Pałka and Lippi 2020), the focus has largely rested on the identification of clauses that might be of interest to consumers. The perspective of UOKiK, which is necessary if they are to be supported in processing masses of documents actively, is largely absent. ...

Big data analytics, online terms of service and privacy policies
  • Citing Chapter
  • May 2021

... Through body postures, individuals can convey discrete but precise signals to the robot to execute predefined tasks. Vision-based and wearable-based approaches dominate this field when it comes to creating posture-based interaction interfaces [6]. ...

A General Pipeline for Online Gesture Recognition in Human–Robot Interaction

IEEE Transactions on Human-Machine Systems

... For the method to be successfully applied, it must be accompanied by another method or technique -e.g., 5S or Poka-Yoke The Poka-Yoke method comes from the Toyota company, which was developed by Shigeo Shingo. The method's goal is to prevent human errors by designing process limitations or eliminating quality control (Martinelli et al., 2022) using anything that can detect errors that can reduce the quality of the end product (Santos et al., 2023). Common devices that are used in Poka-Yoke systems are flashing lights, alarms, sensors, and sliding rails (Wijaya et al., 2020). ...

Poka Yoke Meets Deep Learning: A Proof of Concept for an Assembly Line Application

Applied Sciences

... Digital twins are also being referred to as "living models" that continually adapt to changes in the environment or operation using realtime sensory data and can forecast the future of the corresponding physical assets [40] [33]. Therefore a way to introduce individual and collective self-development capabilities [36] [37] is required. ...

Individual and Collective Self-Development: Concepts and Challenges

... Hence, our study started with the development of a retrieval and ranking tool based on MARGOT, called AMICA (13). AMICA can automatically process scientific articles, and identify features that are relevant to a key phrase given in input (user query), for example, a sequence of keywords linked to a particular pathology. ...

AMICA: An Argumentative Search Engine for COVID-19 Literature
  • Citing Conference Paper
  • July 2022

Marco Lippi

·

·

·

[...]

·