PreprintPDF Available

Practical Applications of Fed_LLM for Privacy- Preserving AI in Education

Authors:
Preprints and early-stage research may not have been peer reviewed yet.
A.1 Practical Applications of Fed_LLM for Privacy-
Preserving AI in Education
Sandra Hummel1, Gitanjali Wadhwa1, Syed Hur Abbas1,
Mana-Teresa Donner1
1 TUD Dresden University of Technology, ScaDS.AI Dresden
1 Introduction
The Fed_LLM framework integrates federated learning and large language models
(LLMs) to create a cutting-edge educational system. This innovative approach
optimizes learning experiences and ensures robust data privacy, making it an
essential tool for modern education. Fed_LLM is practically applied and tested in
educational settings, focusing on its products, concepts, and real-world
implementation.
In educational settings, the demand for personalized, efficient, and secure learning
tools is at an all-time high (Hummel et al., 2023; Hummel & Donner, 2023; Bartok
et al., 2023). Traditional educational systems often struggle to meet these needs
(Egger & Hummel, 2020), especially when it comes to maintaining data privacy
while providing tailored learning experiences. The Fed_LLM framework addresses
these challenges by combining the decentralized nature of federated learning with
the advanced capabilities of LLMs. This combination not only personalizes
educational content to cater to individual student needs but also ensures that
sensitive data remains protected on local devices (Li et al., 2020; Yang et al., 2019).
The impressive performance of generative pre-trained LLMs and their multi-modal
derivations largely owes to their capacity to learn representations at scale. However,
this capability comes with significant energy and monetary costs due to the
extensive computational resources required for training these models (Hoffmann et
al., 2022). Traditionally, training large LLMs involves thousands of hours of
computation on specialized hardware in centralized data centers, a process that
incurs high energy and monetary costs. Distributing training across multiple data
centers can further escalate these costs due to communication overheads (Lin et al.,
2020). Federated learning (FL) presents a promising solution by enabling
8
professional development programs should be established to assist educators in
integrating Fed_LLM into their pedagogy.
Although federated learning addresses many privacy concerns by keeping data
localized, continuous monitoring and robust cybersecurity measures are necessary
to protect student data from potential breaches. The integration of Fed_LLM into
educational practice must be pedagogically sound, aligning with educational goals
and enhancing the learning process. Research should explore the most effective
ways to integrate Fed_LLM into various educational settings to maximize its
benefits. Implementing Fed_LLM involves significant financial and resource
investments, so educational institutions must carefully consider their budgets and
allocate resources effectively to ensure the sustainability of the technology.
Partnerships with technology companies and government funding could help
alleviate some financial burdens.
The ethical implications of using AI in education must also be thoroughly examined,
including addressing potential biases in AI algorithms and ensuring that the
technology promotes fairness and inclusivity. By addressing these critical
perspectives and leveraging the innovative capabilities of Fed_LLM, educational
institutions can significantly enhance the quality of education, providing a more
personalized, efficient, and secure learning environment for all students. This
comprehensive approach ensures that the transformative potential of Fed_LLM is
fully realized, marking a significant step forward in the advancement of educational
technology.
References
Bartok, L., Donner, M.-T., Ebner, M., Gosch, N., Handle-Pfeiffer, D., Hummel, S.,
Kriegler-Kastelic, G., Leitner, P., Tang, T., Veljanova, H., Winter, C., &
Zwiauer, C. (2023). Learning Analytics Studierende im Fokus. Zeitschrift
für Hochschulentwicklung: ZFHE; Beiträge zu Studium, Wissenschaft und
Beruf, 18 (Sonderheft Hochschullehre), 223-250.
https://doi.org/10.3217/zfhe-SH-HL/12
Bohlinger, S., & Hummel, S. (2024). Digital Capacity Building in Teacher
Education: An Environmental Case Study from Cambodia. In S. Hummel
(Ed.), Empowering Education in Cambodia and Sri Lanka: Advancing
9
Quality in 21st Century Teaching and Learning (pp. 31-50). Springer.
https://doi.org/10.1007/978-3-658-43306-2_3
Carlini, N., Liu, C., Erlingsson, U., Kos, J., & Song, D. (2019). The Secret Sharer:
Evaluating and Testing Unintended Memorization in Neural Networks. In
28th USENIX Security Symposium (USENIX Security 19) (pp. 267-284).
Santa Clara, CA: USENIX Association.
https://www.usenix.org/conference/usenixsecurity19/presentation/carlini
Egger, R., & Hummel, S. (2020). Stolperstein oder Kompetenzstufe. Die
Studieneingangsphase und ihre Bedeutung für die Wissenschaftssozialisation
von Studierenden. Springer VS. https://doi.org/10.1007/978-3-658-23283-2
Geyer, R. C., Klein, T., & Nabi, M. (2017). Differentially Private Federated
Learning: A Client-Level Perspective. arXiv preprint arXiv:1712.07557.
https://arxiv.org/abs/1712.07557
Hoffmann, J., Borgeaud, S., Mensch, A., Buchatskaya, E., Cai, T., Rutherford, E.,
Millican, K., van den Driessche, G., Lespiau, J. B., Bennett, L., Schmid, L.,
Scicluna, M., Zimermann, T., Rae, J. W., Veličković, P., Ring, R.,
Babuschkin, I., Martens, L., ... Irving, G. (2022). Training Compute-Optimal
Large Language Models. arXiv preprint arXiv:2203.15556.
https://arxiv.org/abs/2203.15556
Hummel, S., Donner, M.-T., & Egger, R. (2023). Lehren mit Learning Analytics.
Qualifizierungsprogramm für die Hochschullehre. Universität Graz.
https://oer-portal.uni-graz.at/edu-
sharing/components/collections?viewType=1&id=df2fc7c7-b832-421b-
acec-764181c6c2f0
Hummel, S. (2020). Grundlagen der Hochschullehre. Springer.
https://doi.org/10.1007/978-3-658-28181-6
Hummel, S., Egger, R., & Donner, M.-T. (2023). Lehren mit KI: Learning Analytics
für mehr Studierendenorientierung in der Hochschullehre? In T. Köhler, E.
Schopp, N. Kahnwald & R. Sonntag (Eds.), Communities in New Media.
Inclusive digital: Forming Community in an Open Way Self-determined
Participation in the Digital Transformation: Proceedings of 26th Conference
GeNeMe (pp. 7-16). TUDpress. https://doi.org/10.25368/2024.7
Hummel, S., & Donner, M.-T. (2023). KI-Anwendungen in der Hochschulbildung
aus Studierendenperspektive. FNMA Magazin, 3, 38-41.
https://doi.org/10.13140/RG.2.2.13967.97446
Hummel, S., Aldrian, S., & Sheehan, B. (2024). Culture, Education and Digital
Transformation in Sri Lanka. In S. Hummel (Ed.), Empowering Education in
Cambodia and Sri Lanka: Advancing Quality in 21st Century Teaching and
Learning (pp. 9-29). Springer. https://doi.org/10.1007/978-3-658-43306-2_2
Hummel, S., Brodacz-Geier, M., & Sheehan, B. (2024). Pedagogical
Professionalization in Digital Contexts. In S. Hummel (Ed.), Empowering
10
Education in Cambodia and Sri Lanka: Advancing Quality in 21st Century
Teaching and Learning (pp. 203-219). Springer. https://doi.org/10.1007/978-
3-658-43306-2_13
Hummel, S., Bohlinger, S., & Sheehan, B. (2024). Modern Pedagogies in the Digital
Era: Elevating Educational Standards in Teacher Education. In S. Hummel
(Ed.), Empowering Education in Cambodia and Sri Lanka: Advancing
Quality in 21st Century Teaching and Learning (pp. 65-76). Springer.
https://doi.org/10.1007/978-3-658-43306-2_5
Kairouz, P., McMahan, H. B., Avent, B., Bellet, A., Bennis, M., Bhagoji, A. N., ...
& Zhao, S. (2021). Advances and Open Problems in Federated Learning.
Foundations and Trends® in Machine Learning, 14(1-2), 1-210.
https://doi.org/10.1561/2200000083
Li, T., Sahu, A. K., Talwalkar, A., & Smith, V. (2020). Federated Learning:
Challenges, Methods, and Future Directions. IEEE Signal Processing
Magazine, 37(3), 50-60. https://doi.org/10.1109/MSP.2020.2975749
McMahan, H. B., Moore, E., Ramage, D., Hampson, S., & y Arcas, B. A. (2017).
Communication-Efficient Learning of Deep Networks from Decentralized
Data. In Artificial Intelligence and Statistics (pp. 1273-1282). PMLR.
https://arxiv.org/abs/1602.05629
Rieke, N., Hancox, J., Li, W., Milletari, F., Roth, H. R., Albarqouni, S., ... & Bakas,
S. (2020). The Future of Digital Health with Federated Learning. NPJ Digital
Medicine, 3(1), 1-7. https://doi.org/10.1038/s41746-020-00323-1
Yang, Q., Liu, Y., Chen, T., & Tong, Y. (2019). Federated Machine Learning:
Concept and Applications. ACM Transactions on Intelligent Systems and
Technology (TIST), 10(2), 1-19. https://doi.org/10.1145/3298981
ResearchGate has not been able to resolve any citations for this publication.
Chapter
Full-text available
This article discusses the role of the EU-funded project ‘CONTESSA – Contemporary teaching skills’ in modernizing teacher training methodologies in the two Asian countries Cambodia and Sri Lanka. Focusing on the pressing need to adapt to 21st-century teaching and learning paradigms, CONTESSA's design reflects an understanding of the technological disparities across these countries. While acknowledging the resource constraints, the program centers its e-learning curriculum on empowering teachers to enhance lesson planning and teaching quality. The program introduces multiple modules, each targeting specific pedagogical skills ranging from digital teaching, learner-centric approaches, embracing classroom diversity, to understanding gender-based nuances in education. Additionally, the article underscores the significance of training teacher trainers to ensure the program's sustainability, detailing a comprehensive Train-the-Trainer initiative that encompasses face-to-face workshops, tutorials, and a user manual. Derived from a four-year-long project, findings indicate the potential of the CONTESSA program in fostering a shift towards widely accessible teacher training, aiming to equip educators with crucial 21st-century teaching skills and promote holistic student learning. The program's broader vision emphasizes a collaborative community of practice, ensuring the professional growth of educators and impactful learning experiences for students.
Article
Full-text available
Data-driven machine learning (ML) has emerged as a promising approach for building accurate and robust statistical models from medical data, which is collected in huge volumes by modern healthcare systems. Existing medical data is not fully exploited by ML primarily because it sits in data silos and privacy concerns restrict access to this data. However, without access to sufficient data, ML will be prevented from reaching its full potential and, ultimately, from making the transition from research to clinical practice. This paper considers key factors contributing to this issue, explores how federated learning (FL) may provide a solution for the future of digital health and highlights the challenges and considerations that need to be addressed.
Article
Full-text available
Federated learning (FL) is a machine learning setting where many clients (e.g. mobile devices or whole organizations) collaboratively train a model under the orchestration of a central server (e.g. service provider), while keeping the training data decentralized. FL embodies the principles of focused data collection and minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, centralized machine learning and data science approaches. Motivated by the explosive growth in FL research, this paper discusses recent advances and presents an extensive collection of open problems and challenges. * Peter Kairouz and H. Brendan McMahan conceived, coordinated, and edited this work. Correspondence to kairouz@ google.com and mcmahan@google.com.
Article
Federated learning involves training statistical models over remote devices or siloed data centers, such as mobile phones or hospitals, while keeping data localized. Training in heterogeneous and potentially massive networks introduces novel challenges that require a fundamental departure from standard approaches for large-scale machine learning, distributed optimization, and privacy-preserving data analysis. In this article, we discuss the unique characteristics and challenges of federated learning, provide a broad overview of current approaches, and outline several directions of future work that are relevant to a wide range of research communities.
Article
Today’s artificial intelligence still faces two major challenges. One is that, in most industries, data exists in the form of isolated islands. The other is the strengthening of data privacy and security. We propose a possible solution to these challenges: secure federated learning. Beyond the federated-learning framework first proposed by Google in 2016, we introduce a comprehensive secure federated-learning framework, which includes horizontal federated learning, vertical federated learning, and federated transfer learning. We provide definitions, architectures, and applications for the federated-learning framework, and provide a comprehensive survey of existing works on this subject. In addition, we propose building data networks among organizations based on federated mechanisms as an effective solution to allowing knowledge to be shared without compromising user privacy.
The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks
  • N Carlini
  • C Liu
  • U Erlingsson
  • J Kos
  • D Song
Carlini, N., Liu, C., Erlingsson, U., Kos, J., & Song, D. (2019). The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks. In 28th USENIX Security Symposium (USENIX Security 19) (pp. 267-284). Santa Clara, CA: USENIX Association.
  • R C Geyer
  • T Klein
  • M Nabi
Geyer, R. C., Klein, T., & Nabi, M. (2017). Differentially Private Federated Learning: A Client-Level Perspective. arXiv preprint arXiv:1712.07557. https://arxiv.org/abs/1712.07557
  • J Hoffmann
  • S Borgeaud
  • A Mensch
  • E Buchatskaya
  • T Cai
  • E Rutherford
  • K Millican
  • G Van Den Driessche
  • J B Lespiau
  • L Bennett
  • L Schmid
  • M Scicluna
  • T Zimermann
  • J W Rae
  • P Veličković
  • R Ring
  • I Babuschkin
  • L Martens
  • . . Irving
Hoffmann, J., Borgeaud, S., Mensch, A., Buchatskaya, E., Cai, T., Rutherford, E., Millican, K., van den Driessche, G., Lespiau, J. B., Bennett, L., Schmid, L., Scicluna, M., Zimermann, T., Rae, J. W., Veličković, P., Ring, R., Babuschkin, I., Martens, L.,... Irving, G. (2022). Training Compute-Optimal Large Language Models. arXiv preprint arXiv:2203.15556. https://arxiv.org/abs/2203.15556
A Decentralized Federated Learning Framework via Data-to-Model Transmission
  • Y Lu
  • F Zhuang
  • Q He
Lu, Y., Zhuang, F., & He, Q. (2020). A Decentralized Federated Learning Framework via Data-to-Model Transmission. Neural Computing and Applications, 32(1), 21-33. https://doi.org/10.1007/s00521-019-04035-0