- Joined
- Jan 14, 2019
When did they all become such pussies?
My relatives are old enough to remember the French and the Americans in Vietnam; and most European states have long histories of subjugating other parts of the world. They were not the only people to do this, and both conquerors and slaves came from many races, creeds and cultures.
It is uniquely the white races that seem to have lost this fire however. The Chinese are still ruled by a dynasty and sat upon the wheezing bodies of smaller client states around them. Africa has declined terribly since the white men left and the end of South Africa really was the nail in the coffin for any hope of native development not bought and directed by Israel or China. South America is still the coked up irrelevant shithole it always was too, but all of these people still retain some sense of national pride and determination to rule and govern themselves.
What the fuck happend to you people though? Less than a hundred years ago you were shooting the savages before they got within thirty feet; now you throw yourselves at their feet grovelling for the sin of actually existing.
I'm not aiming this at a particular state in particular. America, Britain, Germany, Canada. Even my own native France is just as guilty.
I see the backward shithole half of my family came from, we realised the way up was through the western way. But now I see the natives willingly dismantling everything to better resemble the desolate wastelands these zealots are so determined to bring here.
I can see the mad on the internet ratings flooding in but I really just can't grasp it. What the fuck is wrong with white people? Why are you all collectively punishing yourselves stead of telling the woke, the muslims or whatever invader it is that week to fall in line or fuck off?
The angriest and most determined anti woke, Islam etc people I know in real life and online? Most of them aren't white. But if I had a cent for every white insta thot spewing this stuff I'd be in Forbes rich list now.
My relatives are old enough to remember the French and the Americans in Vietnam; and most European states have long histories of subjugating other parts of the world. They were not the only people to do this, and both conquerors and slaves came from many races, creeds and cultures.
It is uniquely the white races that seem to have lost this fire however. The Chinese are still ruled by a dynasty and sat upon the wheezing bodies of smaller client states around them. Africa has declined terribly since the white men left and the end of South Africa really was the nail in the coffin for any hope of native development not bought and directed by Israel or China. South America is still the coked up irrelevant shithole it always was too, but all of these people still retain some sense of national pride and determination to rule and govern themselves.
What the fuck happend to you people though? Less than a hundred years ago you were shooting the savages before they got within thirty feet; now you throw yourselves at their feet grovelling for the sin of actually existing.
I'm not aiming this at a particular state in particular. America, Britain, Germany, Canada. Even my own native France is just as guilty.
I see the backward shithole half of my family came from, we realised the way up was through the western way. But now I see the natives willingly dismantling everything to better resemble the desolate wastelands these zealots are so determined to bring here.
I can see the mad on the internet ratings flooding in but I really just can't grasp it. What the fuck is wrong with white people? Why are you all collectively punishing yourselves stead of telling the woke, the muslims or whatever invader it is that week to fall in line or fuck off?
The angriest and most determined anti woke, Islam etc people I know in real life and online? Most of them aren't white. But if I had a cent for every white insta thot spewing this stuff I'd be in Forbes rich list now.
Last edited: