Skip to content

Commit

Permalink
delete test
Browse files Browse the repository at this point in the history
Signed-off-by: wyoung1 <[email protected]>
  • Loading branch information
wyoung1 committed Sep 25, 2024
1 parent a0f733b commit 14801f1
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 16 deletions.
2 changes: 1 addition & 1 deletion examples/imagenet/multiedge_inference_bench/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ You can view the graphical representation of relevant metrics in /ianvs/multiedg
To compare the running conditions of the model with and without parallelism in the multiedge inference scenario, you can modify the value of --devices_info in base_model.py to devices_one.yaml to view the relevant metrics when the model runs on a single device.

## Step 5. Run Benchmarking Job - Automatic
We offer a profiling-based plus memory matching segmentation algorithm to compare with the method of manually specifying partitioning points. This method prioritizes the memory matching between the computational subgraph and the device. First, we profile the initial model on the CPU to collect memory usage, the number of parameters, computational cost, and the input and output data shapes for each layer, as well as the total number of layers and their names in the entire model. To facilitate subsequent integration, we have implemented profiling for three types of transformer models: vit, bert, and deit. Secondly, based on the results of the profiling and the device information provided in devices.yaml, we can identify the partitioning point that matches the device memory through a single traversal and perform model partitioning.
We offer a profiling-based and memory matching partition algorithm to compare with the method of manually specifying partitioning points. This method prioritizes the memory matching between the computational subgraph and the device. First, we profile the initial model on the CPU to collect memory usage, the number of parameters, computational cost, and the input and output data shapes for each layer, as well as the total number of layers and their names in the entire model. To facilitate subsequent integration, we have implemented profiling for three types of transformer models: vit, bert, and deit. Secondly, based on the results of the profiling and the device information provided in devices.yaml, we can identify the partitioning point that matches the device memory through a single traversal and perform model partitioning.

You should first run the following command to generate a profiling result:
```shell
Expand Down

This file was deleted.

0 comments on commit 14801f1

Please sign in to comment.