Was wondering if there’s anywhere non Islamic that still treats men as king? Where women don’t rule and control everything.
Are there any cities or states in America that men are still in charge of?
If not perhaps we should form our own community or town where we can stick together? No homo. There’d be women just women who have been taught to shave their armpits.