Item Infomation
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Haoran, Wang | - |
dc.contributor.author | Thibaut, Tachon | - |
dc.contributor.author | Chong, Li | - |
dc.date.accessioned | 2023-04-24T02:02:17Z | - |
dc.date.available | 2023-04-24T02:02:17Z | - |
dc.date.issued | 2022 | - |
dc.identifier.uri | https://link.springer.com/article/10.1007/s10766-022-00741-6 | - |
dc.identifier.uri | https://dlib.phenikaa-uni.edu.vn/handle/PNK/8234 | - |
dc.description | CC BY | vi |
dc.description.abstract | The increasing size of deep neural networks (DNNs) raises a high demand for distributed training. An expert could find good hybrid parallelism strategies, but designing suitable strategies is time and labor-consuming. Therefore, automating parallelism strategy generation is crucial and desirable for DNN designers. Some automatic searching approaches have recently been studied to free the experts from the heavy parallel strategy conception. However, these approaches all rely on a numerical cost model, which requires heavy profiling results that lack portability. These profiling-based approaches cannot lighten the strategy generation work due to the non-reusable profiling value. Our intuition is that there is no need to estimate the actual execution time of the distributed training but to compare the relative cost of different strategies. | vi |
dc.language.iso | en | vi |
dc.publisher | Springer | vi |
dc.subject | DNNs | vi |
dc.subject | DNN designers | vi |
dc.title | SMSG: Profiling-Free Parallelism Modeling for Distributed Training of DNN | vi |
dc.type | Book | vi |
Appears in Collections | ||
OER - Công nghệ thông tin |
Files in This Item: