In a country as saturated in Christian assumptions as the United States, there could be no escaping their influence—even for those who imagined that they had. America’s culture wars were less a war against Christianity than a civil war between Christian factions.