TIST 2011
PLDA+
Abstract
Previous methods of distributed Gibbs sampling for LDA run into either memory or communication bottlenecks. To improve scalability, we propose four strategies: data placement, pipeline processing, word bundling, and priority-based scheduling. Experiments show that our strategies significantly reduce the unparallelizable communication bottleneck and achieve good load balancing, and hence improve scalability of LDA.
Authors
Keywords
No keywords are indexed for this paper.
Context
- Venue
- ACM Transactions on Intelligent Systems and Technology
- Archive span
- 2010-2026
- Indexed papers
- 1415
- Paper id
- 1084234275243846381