Companies in my city now advertise to charge women 20% less because of gender paygap. Things that were once rediculous become normal now.
My sister asks me now where all the good men are after shes finished with her feminist rants of why she does not find a job with her liberal arts degree. Shes 32 now and I believe she will never have children in her life. Sadly I think thats for the best.
Random white knight strangers called me "the last sexist of the university" in public after I asked my friend why she assumes that women and men have the same productivity in all careers.
Feminist bullshit literally destroys lifes. And no one dares to do anything against it. I can watch how my once beautiful country slowly drifts into feminist totalitarianism.
Most men become weak and support the feminine imperative. And the strong ones "just enjoy the decline". The only glimpse of hope is that the red pill becomes more popular even among normies.
So what can I do? Do I have to become even more depressed every day until I leave my home country? Or do we have to wait until incels stage a coup and call out the "the fourth Reich"?