Was all this nationalistic talk from the American evangelical Church just that—talk? Or was it indicative of a serious effort to restructure the relationship between the state and the country’s dominant religion? And if it was the latter, why weren’t sane Christians doing more to stop it?