The effort to do so is happening every day in America and other Western nations. It's not a conspiracy theory to think that powerful elites are working tirelessly to eliminate traditional people and culture in the West when they openly admit it and celebrate it. Proof is in their actions and it's all around you. You only need to pay attention and understand it. It's called genocide.
Can I point out again the importance of propaganda in Hollywood media in their efforts to destroy the West (see video)? It's interesting Hollywood never sides with traditional American values or Christian values. Hollywood always takes the side of undermining traditional social order.