The West After the "good guys" Won WW2 - Current

6 months ago
227

We are told the "good guys" won in WW2, so what has happened to Europe and America since then and why? We have lived through a moral and racial degeneration since then in slow motion. The right blames the left and the left blames the right. BOTH are controlled by the same masters, the "victims" of WW2 and the "evil Nazis". Our history has been subverted and inverted by the same human looking rats. Time to wake up, Goy!

**FAIR USE**

Copyright Disclaimer under section 107 of the Copyright Act 1976, allowance is made for “fair use” for purposes such as criticism, comment, news reporting, teaching, scholarship, education and research.

Fair use is a use permitted by copyright statute that might otherwise be infringing.

Non-profit, educational or personal use tips the balance in favor of fair use.

Loading comments...