
Has the Decline of U.S. Christianity Finally Stopped?
“When will American Christianity finally hit rock bottom?” Many church leaders have had that question on their minds as they’ve watched affiliation trends for the past decade. Now, according to Pew Research Center’s recently released Religious Landscape …