Is America a Christian Nation?


Quick Answer

Many of the founders of the United States were Christian, but many were Deist, a faith tradition that believes in God but rejects Christian beliefs such as the divinity of Jesus. The U.S. Constitution was designed to prohibit religious tests for public office and forbid laws respecting certain religious traditions.

Continue Reading
Related Videos

Full Answer

The Constitution's First Amendment ensures what is often referred to as "freedom of religion."

The definition of "Christian nation" is contested by many political and religious thinkers. The Treaty of Tripoli, written in 1797 during President John Adams' administration, explicitly states that the United States wasn't founded on the Christian religion. However, many Christian and right-wing scholars claim that the strong presence of Christian religious beliefs makes America a "Christian nation" regardless of the founding father's intent.

Learn more about Christianity

Related Questions