https://github.com/ds2-lab/wukong
Wukong: A scalable and locality-enhanced serverless parallel framework (ACM SoCC'20)
https://github.com/ds2-lab/wukong
analytics aws aws-lambda cloud-computing dask data-analytics faas linear-algebra machine-learning parallel-computing python serverless serverless-computing
Last synced: 8 months ago
JSON representation
Wukong: A scalable and locality-enhanced serverless parallel framework (ACM SoCC'20)
- Host: GitHub
- URL: https://github.com/ds2-lab/wukong
- Owner: ds2-lab
- Created: 2019-10-05T00:55:13.000Z (over 6 years ago)
- Default Branch: socc2020
- Last Pushed: 2024-11-02T21:46:49.000Z (over 1 year ago)
- Last Synced: 2025-05-19T18:11:25.187Z (10 months ago)
- Topics: analytics, aws, aws-lambda, cloud-computing, dask, data-analytics, faas, linear-algebra, machine-learning, parallel-computing, python, serverless, serverless-computing
- Language: Python
- Homepage: https://ds2-lab.github.io/Wukong/
- Size: 14.6 MB
- Stars: 74
- Watchers: 6
- Forks: 16
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
README
# Wukong

[](https://leap-wukong.readthedocs.io/en/latest/?badge=latest)
Wukong is a high-performance and highly scalable locality-aware, serverless workflow and DAG engine. Wukong uses serverless computing to accelerate the execution of DAG-based scientific, linear algebra, machine learning, and data analytics workloads.
## What is Wukong?
Wukong is a serverless parallel computing framework attuned to FaaS platforms such as AWS Lambda. Wukong provides decentralized scheduling using a combination of static and dynamic scheduling. Wukong supports general Python data analytics workloads at any scale.

## Publications
First Paper: In Search of a Fast and Efficient Serverless DAG Engine (Appeared at PDSW '19)
https://arxiv.org/abs/1910.05896
Latest Paper: Wukong: A Scalable and Locality-Enhanced Framework for Serverless Parallel Computing (Appeared at ACM SoCC '20) https://arxiv.org/abs/2010.07268 .
If you use our source code for a publication or project, please cite the paper using this [bibtex](#to-cite-wukong).
This branch contains the source code of Wukong corresponding to the SoCC 2020 publication, which is a later version than the PDSW paper referenced above.
## Installation
**Required Python version**: Python3.12
A majority of the required AWS infrastructure can be created using the provided `aws_setup.py` script in `Wukong/Static Scheduler/install/` directory. Please be sure to read through the `wukong_setup_config.yaml` configuration file located in the same directory prior to running the script. In particular, your public IP address should be added to the configuration file if you'd like SSH to be enabled from your machine to VMs created in the Wukong VPC.
In addition, there is documentation in the `setup/` directory for additional/supplementary instructions concerning the creation of the AWS infrastructure for Wukong.
### Verifying Your Installation
There is a sample Redis configuration file available at `Static Scheduler/install/redis.conf`.
Similarly, there is a simple test application available at `Static Scheduler/simple-test-app.py`. You can use these to test your installation.
Start the proxy by navigating to the `KV Store Proxy` directory and executing the command:
``` sh
python3.12 proxy.py --redis 127.0.0.1
```
You'll also need to start Redis on your machine. In another terminal window, execute the command:
``` sh
redis-server
```
You can use the provided sample Redis configuration file (`Static Scheduler/install/redis.conf`).
Finally, to test your installation, you may run:
``` sh
python3.12 simple-test-app.py
```
from the `Static Scheduler` directory.
If your installation is working, then you should see, among other things, the following in the program's output:
```
...
[CLIENT] Obtained value for key incr- from Redis.
[CLIENT] Returning {'status': 'OK', 'data': {'incr-': 6}} to the user...
Result: 6
...
```
## Code Overview/Explanation
*This section is currently under development...*
### The Static Scheduler
Generally speaking, a user submits a job by calling the `.compute()` function on an underlying Dask collection. Support for Dask's asynchronous `client.compute()` API is coming soon.
When the `.compute()` function is called, the `update_graph()` function is called within the static Scheduler, specifically in **scheduler.py**. This function is responsible for adding computations to the Scheduler's internal graph. It's triggered whenever a Client calls `.submit()`, `.map()`, `.get()`, or `.compute()`. The depth-first search (DFS) method is defined with the `update_graph` function, and the DFS also occurs during the `update_graph` function's execution.
Once the DFS has completed, the Scheduler will serialize all of the generated paths and store them in the KV Store (Redis). Next, the Scheduler will begin the computation by submitting the leaf tasks to the `BatchedLambdaInvoker` object (which is defined in **batched_lambda_invoker.py**). The "Leaf Task Invoker" processes are defined within the `BatchedLambdaInvoker` class as the `invoker_polling_process` function. Additionally, the `_background_send` function is running asynchronously on an interval (using [Tornado](https://www.tornadoweb.org/en/stable/)). This function takes whatever tasks have been submitted by the Scheduler and divdes them up among itself and the Leaf Task Invoker processes, which then invoke the leaf tasks.
The Scheduler listens for results from Lambda using a "Subscriber Process", which is defined by the `poll_redis_process` function. This process is created in the Scheduler's `start` function. (All of this is defined in **scheduler.py**.) The Scheduler is also executing the `consume_redis_queue()` function asynchronously (i.e., on the [Tornado IOLoop](https://www.tornadoweb.org/en/stable/ioloop.html)). This function processes whatever messages were received by the aforementioned "Subscriber Process(es)". Whenever a message is processed, it is passed to the `result_from_lambda()` function, which begins the process of recording the fact that a "final result" is available.
### The KV Store Proxy
This component is used to parallelize Lambda function invocations in the middle of a workload's execution. This is advantageous, as invoking individual AWS Lambda functions has a relatively high overhead (~50ms per invocation). For large fan-outs, this overhead can bottleneck Wukong's performance. Thus, the KV store proxy alleviates that by leveraging a VM with many cores to parallelize such invocations.
### The AWS Lambda Task Executor
The Task Executors are responsible for executing tasks and performing dynamic scheduling. Executors cooperate with one another to decide who should execute downstream tasks during fan-ins. Executors communicate through intermediate storage (e.g., Redis).
### Developer Setup Notes
When setting up Wukong, make sure to update the variables referencing the name of the AWS Lambda function used as the Wukong Task Executor. For example, in "AWS Lambda Task Executor/function.py", this is a variable *lambda_function_name* whose value should be the same as the name of the Lambda function as defined in AWS Lambda itself.
There is also a variable referencing the function's name in "Static Scheduler/wukong/batched_lambda_invoker.py" (as a keyword argument to the constructor of the BatchedLambdaInvoker object) and in "KV Store Proxy/proxy_lambda_invoker.py" (also as a keyword argument to the constructor of ProxyLambdaInvoker).
By default, Wukong is configured to run within the us-east-1 region. If you would like to use a different region, then you need to pass the "region_name" parameter to the Lambda Client objects created in "Static Scheduler/wukong/batched_lambda_invoker.py", "KV Store Proxy/proxy_lambda_invoker.py", "KV Store Proxy/proxy.py", "AWS Lambda Task Executor/function.py", and "Static Scheduler/wukong/scheduler.py".
## Code Examples
In the following examples, modifying the value of the *chunks* parameter will essentially change the granularity of the tasks generated in the DAG. Essentially, *chunks* specifies how the initial input data is partitioned. Increasing the size of *chunks* will yield fewer individual tasks, and each task will operate over a large proportion of the input data. Decreasing the size of *chunks* will result in a greater number of individual tasks, with each task operating on a smaller portion of the input data.
### LocalCluster Overview
```
LocalCluster(object):
host : string
The public DNS IPv4 address associated with the EC2 instance on which the Scheduler process is executing, along with the port on
which the Scheduler is listening. The format of this string should be "IPv4:port".
n_workers : int,
Artifact from Dask. Leave this at zero.
proxy_adderss : string,
The public DNS IPv4 address associated with the EC2 instance on which the KV Store Proxy process is executing.
proxy_port : 8989,
The port on which the KV Store Proxy process is listening.
redis_endpoints : list of tuples of the form (string, int)
List of the public DNS IPv4 addresses and ports on which KV Store (Redis) instances are listening. The format
of this list should be [("IP_1", port_1), ("IP_2", port_2), ..., ("IP_n", port_n)]
num_lambda_invokers : int
This value specifies how many 'Initial Task Executor Invokers' should be created by the Scheduler. The 'Initial Task
Executor Invokers' are processes that are used by the Scheduler to parallelize the invocation of Task Executors
associated with leaf tasks. These are particularly useful for large workloads with a big number of leaf tasks.
max_task_fanout : int
This specifies the size of a "fanout" required for a Task Executor to utilize the KV Store Proxy for parallelizing downstream
task invocation. The principle here is the same as with the initial task invokers. Our tests found that invoking Lambda functions
takes about 50ms on average. As a result, if a given Task T has a large fanout (i.e., there are a large number of downstream tasks
directly dependent on T), then it may be advantageous to parallelize the invocation of these downstream tasks.
use_fargate : bool
If True, then Wukong will attempt to use AWS Fargate for its intermediate storage. This requires that the AWS Fargate infrastructure
already exists and that Wukong has been correctly configured to use it (i.e., passing the required information to the 'LocalCluster'
instance. This defaults to False, in which case Wukong simply uses a single Redis instance for all intermediate storage.
use_local_proxy : bool
If True, automatically deploy the KV Store Proxy locally on the same VM as the Static Scheduler.
Note that the user should pass a value for the `local_proxy_path` property if setting `use_local_proxy` to True.
If not, then Wukong will attempt to locate the proxy in "../KV Store Proxy/", which may or may not work
depending on where Wukong is being executed from.
local_proxy_path: str
Fully-qualified path to the KV Store Proxy source code, specifically the proxy.py file.
This is only used when `use_local_proxy` is True.
Example: "/home/ec2-user/Wukong/KV Store Proxy/proxy.py"
```
### Single-Node DAG Example
```python
import dask.array as da
from dask import delayed
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
# Define a function.
def incr(x):
return x + 1
example_computation = delayed(incr)(5)
# Start the workload.
result = example_computation.compute(scheduler = client.get)
print("Result: %d" % result)
```
### 3-Node DAG Examples
```python
import dask.array as da
from dask import delayed
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
# Define some functions.
def incr(x):
return x + 1
def decr(x):
return x - 1
def double(x):
return x * 2
def add_values(x, y):
return x + y
# Linear 3-Node DAG
a = delayed(incr)(5)
b = delayed(decr)(a)
c = delayed(double)(b)
result1 = c.compute(scheduler = client.get)
print("Result: %d" % result1)
# 3-Node DAG with a Fan-In
x = delayed(incr)(3)
y = delayed(decr)(7)
z = delayed(add_values)(x,y)
result2 = z.compute(scheduler = client.get)
print("Result: %d" % result2)
```
### Tree Reduction
``` python
from dask import delayed
import operator
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
L = range(1024)
while len(L) > 1:
L = list(map(delayed(operator.add), L[0::2], L[1::2]))
# Start the computation.
L[0].compute(scheduler = client.get)
```
### SVD of 'Tall-and-Skinny' Matrix
```python
import dask.array as da
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
# Compute the SVD of 'Tall-and-Skinny' Matrix
X = da.random.random((200000, 1000), chunks=(10000, 1000))
u, s, v = da.linalg.svd(X)
# Start the computation.
v.compute(scheduler = client.get)
```
### SVD of Square Matrix with Approximation Algorithm
```python
import dask.array as da
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
# Compute the SVD of 'Tall-and-Skinny' Matrix
X = da.random.random((10000, 10000), chunks=(2000, 2000))
u, s, v = da.linalg.svd_compressed(X, k=5)
# Start the computation.
v.compute(scheduler = client.get)
```
### GEMM (Matrix Multiplication)
``` python
import dask.array as da
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
x = da.random.random((10000, 10000), chunks = (1000, 1000))
y = da.random.random((10000, 10000), chunks = (1000, 1000))
z = da.matmul(x, y)
# Start the computation.
z.compute(scheduler = client.get)
```
### Parallelizing Prediction (sklearn.svm.SVC)
``` python
import pandas as pd
import seaborn as sns
import sklearn.datasets
from sklearn.svm import SVC
import dask_ml.datasets
from dask_ml.wrappers import ParallelPostFit
from wukong import LocalCluster, Client
local_cluster = LocalCluster(host=':8786',
proxy_address = '',
num_lambda_invokers = 4,
# Automatically create proxy locally. Pass same IPv4 for `host` and `proxy_address`
use_local_proxy = True,
# Path to `proxy.py` file.
local_proxy_path = "/home/ec2-user/Wukong/KV Store Proxy/proxy.py",
redis_endpoints = [("", 6379)],
use_fargate = False)
client = Client(local_cluster)
X, y = sklearn.datasets.make_classification(n_samples=1000)
clf = ParallelPostFit(SVC(gamma='scale'))
clf.fit(X, y)
X, y = dask_ml.datasets.make_classification(n_samples=800000,
random_state=800000,
chunks=800000 // 20)
# Start the computation.
clf.predict(X).compute(scheduler = client.get)
```
## To Cite Wukong
```
@inproceedings{socc20-wukong,
author = {Carver, Benjamin and Zhang, Jingyuan and Wang, Ao and Anwar, Ali and Wu, Panruo and Cheng, Yue},
title = {Wukong: A Scalable and Locality-Enhanced Framework for Serverless Parallel Computing},
year = {2020},
isbn = {9781450381376},
publisher = {Association for Computing Machinery},
url = {https://doi.org/10.1145/3419111.3421286},
doi = {10.1145/3419111.3421286},
series = {SoCC '20}
}
```
```
@INPROCEEDINGS {pdsw19-wukong,
author = {B. Carver and J. Zhang and A. Wang and Y. Cheng},
booktitle = {2019 IEEE/ACM Fourth International Parallel Data Systems Workshop (PDSW)},
title = {In Search of a Fast and Efficient Serverless DAG Engine},
year = {2019},
doi = {10.1109/PDSW49588.2019.00005},
url = {https://doi.ieeecomputersociety.org/10.1109/PDSW49588.2019.00005},
publisher = {IEEE Computer Society}
}
```