Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/MachineLearningSystem/SC21_Ribbon
https://github.com/MachineLearningSystem/SC21_Ribbon
Last synced: 9 days ago
JSON representation
- Host: GitHub
- URL: https://github.com/MachineLearningSystem/SC21_Ribbon
- Owner: MachineLearningSystem
- License: mit
- Fork: true (boringlee24/SC21_Ribbon)
- Created: 2022-11-15T12:46:29.000Z (about 2 years ago)
- Default Branch: main
- Last Pushed: 2022-05-06T20:41:25.000Z (over 2 years ago)
- Last Synced: 2024-08-02T19:34:25.808Z (4 months ago)
- Size: 987 KB
- Stars: 0
- Watchers: 0
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- awesome-AI-system - RIBBON: cost-effective and qos-aware deep learning model inference using a diverse pool of cloud computing instances SC'21
README
# SC21 RIBBON: cost-effective and qos-aware deep learning model inference using a diverse pool of cloud computing instances
Ribbon applies a Bayesian Optimization (BO) engine for heterogeneous instance serving of ML inference queries. Please see the link for the full paper here: [RIBBON: cost-effective and qos-aware deep learning model inference using a diverse pool of cloud computing instances](https://dl.acm.org/doi/10.1145/3458817.3476168)
## Dependencies
With Python 3.7 ready, the other required packages can be installed with command
```shell
pip install -r requirements.txt
```## Bayesian Optimization Engine Setup
Ribbon uses a modified public open-source BO library from [fmfn](https://github.com/fmfn/BayesianOptimization)
To setup the BO backend, clone the repo, copy the source file over and build the library
```shell
cd / # replace with custom path
git clone https://github.com/fmfn/BayesianOptimization.git
cp Ribbon/bayesian_optimization.py BayesianOptimization/bayes_opt
cp Ribbon/util.py BayesianOptimization/bayes_opt
cd BayesianOptimization
python setup.py build
PYTHONPATH="$PYTHONPATH://BayesianOptimization/build/lib" # make sure python sees this library
export PYTHONPATH
cd //Ribbon
```
## Inference modelsThe source code for evaluated models are in the ```models``` directory. The characterization data of each model on various instances are in the ```characterization``` directory. To verify the characterization data, navigate to the ```models``` directory, follow the instructions to run the benchmarks, and compare the collected logs with data in ```characterization```.
Here are the links to each model implementation.
1. CANDLE (cancer distributed learning environment) Combo model: [link](https://github.com/ECP-CANDLE/Benchmarks/tree/master/Pilot1/Combo)
2. VGG model: [link](https://keras.io/api/applications/vgg/)
3. ResNet model: [link](https://keras.io/api/applications/resnet/)
4. MT-WND (multi-task wide and deep): [link](https://github.com/harvard-acc/DeepRecSys/blob/master/models/multi_task_wnd.py)
5. DIEN (deep interest evolution network): [link](https://github.com/harvard-acc/DeepRecSys/blob/master/models/dien.py)## Start Ribbon
The characterization data is used to evaluate whether a certain configuration meets the target QoS. First extract the zipped file.
```shell
cd characterization
tar -xf logs.tar.gz
cd ../
```Navigate to the BO directory, run Ribbon and all competing schemes
```shell
cd BO/
./all_scheme.sh
```To visualize the comparison, run
```shell
cd visualize
python num_of_samples.py
python explore_cost.py
```After running the visualization scripts, new figures will appear in the ```visualize``` directory. The ```num_of_samples.png``` picture shows the number of samples to find the optimal instance pool for all schemes, the ```explore_cost.png``` picture shows the total cost of exploration for all schemes.
For further inquries, please contact [[email protected]]([email protected])