Came across yet another Christian pundit scratching his head and trying to explain what went wrong with America. How could a culture so grounded in biblical values do such a u-turn to become one of the most secular communities ever, promoting entertainment and sexual freedom above all else? What happened? Where did Christianity fail? As with other pundits, he tries to pinpoint the key