Publication | Closed Access
Heralding the Future of Federated Learning Framework: Architecture, Tools and Future Directions
20
Citations
19
References
2021
Year
Unknown Venue
Artificial IntelligenceEngineeringMachine LearningFederated SystemsExponential GrowthLearning Management SystemData ScienceFuture DirectionsComputing SystemsInternet Of ThingsFederated Learning FrameworkFederated Database SystemData PrivacyLearning AnalyticsComputer ScienceDistributed LearningMobile ComputingPrivacyData SecurityDecentralized Machine LearningDecentralized PrivacyEdge ComputingFederated LearningTensorflow FederatedBig Data
In today's era, the exponential growth of data and its management is a matter of concern. Machine learning has shown its efficacy in multiple application areas. But machine learning on decentralized data was a hectic task since last decade. A novel technology has gained much importance in recent days i.e., federated learning which deals with training on decentralized and distributed data along with preservation of its privacy. Smartphone data being privacy-sensitive is used for locally training a global model which further is aggregated to generate an updated global model which again is distributed among multiple clients. This paper focuses on presenting the efficacy of federated learning by epitomizing an architecture showing the working mechanism of the technology. Further, this paper exhibits an intersection of on-device machine learning, privacy preservation technology and edge computing i.e., federated learning. Also, we have used TensorFlow Federated, an open source platform to simulate federated learning tasks for MNIST and extended MNIST (E-MNIST) datasets. Further, the results contain the loss and accuracy parameters for ten iterations repeated for six optimizer states (Opt <sub xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">st</sub> ) for each dataset. The peak accuracy that we achieved for MNIST and E-MNIST datasets are 0.843 and 0.853 respectively by using federated averaging algorithm. Further, the minimum loss value that we obtained for MNIST and E-MNIST datasets are 0.652 and 0.646 respectively. The execution time for implementing the algorithm for each dataset is presented in a graphical manner. Finally, certain application areas where federated learning technology has aided are scrutinized.
| Year | Citations | |
|---|---|---|
Page 1
Page 1