You know looking back on it the only thing we ever learned about Germany in school was Nazism.
All these years later and Nazism is for some reason, still such a big focus.
Even though there’s a whole culture and history of people before and after the War which is never talked about.
But either way I’ve heard even in Germany they still haven’t moved past it.
In America were at least able to make jokes about it, but in Germany I’ve heard you can’t even do so much as talk freely about the subject.
You aren’t allowed to have your own opinions on the subject unless they are a state approved position. Denying the Holocaust is illegal, and showing any semblance of support for Hitler.
And so I just wonder, why is it still such a sensitive topic?
All these years later, and its as if people still haven’t chosen to let up on it.
At the very least its past the point in time where people ought to be allowed to have discussions on it.
Because if there’s anything history has proven, it’s that Hitler was right.
Because ever since the takeover of media by you guessed it and the subsequent globalization of society, the masses have been brainwashed into supporting cultural marxism which seeks to dismantle the family, our traditions, and the very bedrocks of our society which keep us in place.
I can’t think of a group more intrinsically evil than the one Hitler sought to erase.
They’ve caused so much damage in such short time.
In every sense they’ve done the bidding of the devil and we have thus far failed to keep them in check.