@inproceedings{0edc013da9c44146a579c3d2d388de58,
title = "Work-in-Progress: Cooperative MLP-Mixer Networks Inference on Heterogeneous Edge Devices through Partition and Fusion",
abstract = "As a newly proposed DNN architecture, MLP-Mixer is attracting increasing attention due to its competitive results compared to CNNs and attention-base networks in various tasks. Although MLP-Mixer only contains MLP layers, it still suffers from high communication costs in edge computing scenarios, resulting in long inference time. To improve the inference performance of an MLP-Mixer model on correlated resource-constrained heterogeneous edge devices, this paper proposes a novel partition and fusion method specific for MLP-Mixer layers, which can significantly reduce the communication costs. Experimental results show that, when the number of devices increases from 2 to 6, our partition and fusion method can archive 1.01-1.27x and 1.54-3.12x speedup in scenarios with heterogeneous and homogeneous devices, respectively.",
keywords = "DNN Inference, Edge Intelligence, MLP-Mixer",
author = "Yiming Li and Shouzhen Gu and Mingsong Chen",
note = "Publisher Copyright: {\textcopyright} 2022 IEEE.; 2022 International Conference on Compilers, Architecture, and Synthesis for Embedded Systems, CASES 2022 ; Conference date: 07-10-2022 Through 14-10-2022",
year = "2022",
doi = "10.1109/CASES55004.2022.00021",
language = "英语",
series = "Proceedings - 2022 International Conference on Compilers, Architecture, and Synthesis for Embedded Systems, CASES 2022",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "29--30",
booktitle = "Proceedings - 2022 International Conference on Compilers, Architecture, and Synthesis for Embedded Systems, CASES 2022",
address = "美国",
}