Source: America Was Founded As A Christian Nation By Christians That Believed That The Bible Is The Word Of God – The Washington Standard