I think anywhere you go has pros and cons. Even places that arent horrible yet can and probably will go woke.
I struggle with this too. My city is getting worse. But where to? Maybe a road trip is in order
I do think that life in the country, away from cities, is always better.
Less of everything bad and more of everything good in nature
Somewhere with strong conservative non multicultural views could fight off the global homo Zionist cabal I don’t see that happening here anymore