Have you noticed how few movies show a peaceful, White society in far-off space, or even on Earth or the moon?
It's all wastelands and Stereotypical Post-Apocalyptic Costume Number 387. Lots of leather, though all the animals are supposedly dead, lots of goggles, as if that is the only style of eyewear looted from the vast stores of warehouses left behind after plague/nukes/aliens/apes raze the Earth.
I would like to see movies made by Whites that show a far more likely scenario: White people in space, because meritocracy doesn't shoehorn tokens into places, (I mean, in post-Obama America) with advances in technology as yet unheard of.
Anyone else tired of the movies being, well, all, er globalist-upped?
They use fear to control, similar to how you spook a cow to get it to move somewhere. They see us like animals, so they use fear for everything media related.
Read between the lines in each separate post-apocalyptic movie and you'll see what they're trying to say, such as "(((Global Warming))) will cause this goy!!". They're really fucking bad about hiding their heavy-handed messages.
Caleb Reed
because dystopias make a better story than utopias ??