CuWide: Towards efficient flow-based training for sparse wide models on GPUs (Extended Abstract)
- Conference Paper
In this paper, we propose an efficient GPU-training framework for the large-scale wide models, named cuWide. To fully benefit from the memory hierarchy of GPU, cuWide applies a new flow-based schema for training, which leverages the spatial and temporal locality of wide models to drastically reduce the amount of communication with GPU global memory. Comprehensive experiments show that cuWide can be up to more than 20× faster than the state-of-the-art GPU solutions and multi-core CPU solutions. © 2021 IEEE Show more
Book title2021 IEEE 37th International Conferecne on Data Engineering (ICDE)
Pages / Article No.
MoreShow all metadata