Proceedings of the Fifth ACM International Conference on Web Search and Data Mining 2012
DOI: 10.1145/2124295.2124312
|View full text |Cite
|
Sign up to set email alerts
|

Scalable inference in latent variable models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
220
2

Year Published

2012
2012
2019
2019

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 192 publications
(226 citation statements)
references
References 15 publications
2
220
2
Order By: Relevance
“…Such a setup could be based either on the mcparallel and mccollect functions in the parallel package, which unfortunately are not available on Windows, or by avoiding parallel entirely to use the Rmpi package directly (Yu 2002). Synchronization in Steps 2 and 4 is required to obtain a consistent BN and thus precludes the use of partial update techniques such as that described in Ahmed, Aly, Gonzalez, Narayanamurthy, and Smola (2012).…”
Section: Discussionmentioning
confidence: 99%
“…Such a setup could be based either on the mcparallel and mccollect functions in the parallel package, which unfortunately are not available on Windows, or by avoiding parallel entirely to use the Rmpi package directly (Yu 2002). Synchronization in Steps 2 and 4 is required to obtain a consistent BN and thus precludes the use of partial update techniques such as that described in Ahmed, Aly, Gonzalez, Narayanamurthy, and Smola (2012).…”
Section: Discussionmentioning
confidence: 99%
“…This is relevant since latent variable models and their inference algorithms store and exchange parameters that are associated with vertices rather than edges [1]. Network Topology In many graph-based applications the cost of communication (and to some extent also computation) dwarfs the cost of storing data.…”
Section: Challengesmentioning
confidence: 99%
“…This staleness has two sources: (i) a simple delay due to the asynchronous updates to the local models [3] because a worker computes new gradients without receiving updates from all the other workers; and (ii) a distributed aggregated delay [2] because a worker only completes a mini-batch after it has received all p gradient partitions, requiring multiple synchronisation rounds.…”
Section: Bounding Stalenessmentioning
confidence: 99%
“…A common architecture for DNN systems takes advantage of data-parallelism [3,28]: a set of worker nodes train model replicas on partitions of the input data in parallel; the model replicas are kept synchronised by a set of parameter servers-each server maintains a global partition of the trained model. Periodically workers upload their latest updates to the parameter servers, which aggregate them and return an updated global model.…”
Section: Introductionmentioning
confidence: 99%