dominionism

Noun

 * 1) A tendency among some conservative Christians, especially in the USA, to seek influence or control over secular civil government through political action.
 * 2) The belief that human beings should be free to dominate and exploit nature, including plants and animals.