White women sure be getting uppidity lately. Maybe the problem here isn’t that black people are going further past the line of acceptable behavior and instead that white women are enabling and encouraging it. Seems like this country has really gone to shit since women started becoming “empowered” from the feminist movement. Yes ladies, it’s so empowering to go to work for shit wages (and make it impossible for a single income to sustain a family). The reason women always feel so oppressed isn’t men, it’s always other women I’ve seen it time and again. I got family that are teachers and every single one of them has drama with the other women teachers.
Like you’ll see bullshit on their insta or Facebook like “don’t listen to the haters, they’re just mad that you’re succeeding.” Maybe it’s because I’m a man and don’t have this group of haters and enemies that women all seem to (or they’re batshit crazy and create this fictional force to explain why they aren’t getting to where they think they should be). This country is in the midst of a mental health crisis right now with all these crazy bitches running amok.