Artificial Intelligence-Enabled Smart Learning Environments :Building Adaptive and Personalized Education Systems
Article Sidebar
Main Article Content
With the rapid advancement of machine learning (ML), large-scale data collection has become essential for building accurate models. However, the use of sensitive data introduces significant privacy risks, including data leakage, unauthorized access, and inference attacks. Privacy-Preserving Machine Learning (PPML) has emerged as a crucial research area aimed at enabling data-driven learning while protecting individual privacy. This paper provides a comprehensive overview of major PPML techniques such as homomorphic encryption, differential privacy, secure multi-party computation, and federated learning. It also discusses key challenges including computational overhead, privacy-utility trade-offs, scalability issues, and regulatory concerns. Finally, future research directions are highlighted to guide the development of secure and efficient machine learning systems.
Downloads
References
Kucur, E. N., et al. “Privacy-Preserving Machine Learning Techniques: Cryptographic Approaches…”
MDPI
Xu, R., et al. “Privacy-Preserving Machine Learning: Methods, Challenges and Directions.”
ResearchGate
Parikh, D., et al. “Privacy-Preserving Machine Learning Techniques, Challenges and Research Directions.”

This work is licensed under a Creative Commons Attribution 4.0 International License.
All articles published in our journal are licensed under CC-BY 4.0, which permits authors to retain copyright of their work. This license allows for unrestricted use, sharing, and reproduction of the articles, provided that proper credit is given to the original authors and the source.