Matthew Yeager

54%
Flag icon
One solution, left as an exercise by Pearl in his book on Bayesian networks, is to pretend the graph has no loops and just keep propagating probabilities back and forth until they converge. This is known as loopy belief propagation, both because it works on graphs with loops and because it’s a crazy idea.
The Master Algorithm: How the Quest for the Ultimate Learning Machine Will Remake Our World
Rate this book
Clear rating