Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/APPFL/APPFL
Advanced Privacy-Preserving Federated Learning framework
https://github.com/APPFL/APPFL
federated-learning-framework privacy-preserving-machine-learning
Last synced: 12 days ago
JSON representation
Advanced Privacy-Preserving Federated Learning framework
- Host: GitHub
- URL: https://github.com/APPFL/APPFL
- Owner: APPFL
- License: mit
- Created: 2021-10-07T18:55:33.000Z (about 3 years ago)
- Default Branch: main
- Last Pushed: 2024-10-29T17:39:08.000Z (18 days ago)
- Last Synced: 2024-10-29T18:58:47.219Z (18 days ago)
- Topics: federated-learning-framework, privacy-preserving-machine-learning
- Language: Python
- Homepage: https://appfl.ai
- Size: 63.6 MB
- Stars: 79
- Watchers: 6
- Forks: 17
- Open Issues: 16
-
Metadata Files:
- Readme: README.md
- License: LICENSE
- Citation: CITATION.cff
Awesome Lists containing this project
- awesome-PETs - APPFL - Advanced Privacy-Preserving Federated Learning built my Argonne National Labs aimed at high-performance computing applications. Includes differential-privacy and the ability to incorporate other PETs (Privacy-Preserving Federated Learning / Open-Source Tools)
README
APPFL - Advanced Privacy-Preserving Federated Learning Framework.APPFL, Advanced Privacy-Preserving Federated Learning, is an open-source and highly extensible software framework that allows research communities to implement, test, and validate various ideas related to privacy-preserving federated learning (FL), and deploy real FL experiments easily and safely among distributed clients to train more robust ML models.With this framework, developers and users can easily
* Train any user-defined machine learning model on decentralized data with optional differential privacy and client authentication.
* Simulate various synchronous and asynchronous PPFL algorihtms on high-performance computing (HPC) architecture with MPI.
* Implement customizations in a plug-and-play manner for all aspects of FL, including aggregation algorithms, server scheduling strategies, and client local trainers.[Documentation](http://appfl.rtfd.io/): please check out our documentation for tutorials, users guide, and developers guide.
## Table of Contents
* [Installation](#hammer_and_wrench-installation)
* [Technical Components](#bricks-technical-components)
* [Framework Overview](#bulb-framework-overview)
* [Citation](#page_facing_up-citation)
* [Acknowledgements](#trophy-acknowledgements)## :hammer_and_wrench: Installation
We highly recommend creating a new Conda virtual environment and install the required packages for APPFL.
```bash
conda create -n appfl python=3.8
conda activate appfl
```### User installation
For most users such as data scientists, this simple installation must be sufficient for running the package.
```bash
pip install pip --upgrade
pip install "appfl[examples]"
```If we want to even minimize the installation of package dependencies, we can skip the installation of a few pacakges (e.g., `matplotlib` and `jupyter`):
```bash
pip install "appfl"
```### Developer installation
Code developers and contributors may want to work on the local repositofy.
To set up the development environment,```bash
git clone --single-branch --branch main https://github.com/APPFL/APPFL.git
cd APPFL
pip install -e ".[dev,examples]"
```On Ubuntu:
If the install process failed, you can try:
```bash
sudo apt install libopenmpi-dev,libopenmpi-bin,libopenmpi-doc
```## :bricks: Technical Components
APPFL is primarily composed of the following six technical components
* Aggregator: APPFL supports several popular algorithms to aggregate one or several client local models.
* Scheduler: APPFL supports several synchronous and asynchronous scheduling algorithms at the server-side to deal with different arrival times of client local models.
* Trianer: APPFL supports several client local trainers for various training tasks.
* Privacy: APPFL supprots several global/local differential privacy schemes.
* Communicator: APPFL supports MPI for single-machine/cluster simulation, and gRPC and Globus Compute with authenticator for secure distributed training.
* Compressor: APPFL supports several lossy compressors for model parameters, including [SZ2](https://github.com/szcompressor/SZ), [SZ3](https://github.com/szcompressor/SZ3), [ZFP](https://pypi.org/project/zfpy/), and [SZx](https://github.com/szcompressor/SZx).## :bulb: Framework Overview
In the design of the APPFL framework, we essentially create the server agent and client agent, using the six technical components above as building blocks, to act on behalf of the FL server and clients to conduct FL experiments. For more details, please refer to our [documentation](http://appfl.rtfd.io/).
## :page_facing_up: Citation
If you find APPFL useful for your research or development, please consider citing the following papers:
```
@article{li2024advances,
title={Advances in APPFL: A Comprehensive and Extensible Federated Learning Framework},
author={Li, Zilinghan and He, Shilan and Yang, Ze and Ryu, Minseok and Kim, Kibaek and Madduri, Ravi},
journal={arXiv preprint arXiv:2409.11585},
year={2024}
}@inproceedings{ryu2022appfl,
title={APPFL: open-source software framework for privacy-preserving federated learning},
author={Ryu, Minseok and Kim, Youngdae and Kim, Kibaek and Madduri, Ravi K},
booktitle={2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)},
pages={1074--1083},
year={2022},
organization={IEEE}
}
```## :trophy: Acknowledgements
This material is based upon work supported by the U.S. Department of Energy, Office of Science, under contract number DE-AC02-06CH11357.