"The Words That Created God" is about faith and its role in creating and sustaining human society. This is a new look at faith, from a secular point of view. What do you think? Where does faith come from? What is it anyway? And why is it important? No. Let's be clear: Why it is essential?