@nerdrotic
A Woke Hollywood Revolution
America is indeed a divided nation and Hollywood has certainly played a role in this division. Hollywood, as one of the most influential institutions in the world, has a responsibility to reflect the society in which it exists. However, in recent years, it has been accused of promoting a particular political ideology, one that is often referred to as “woke.” …
Top 5 Woke Hollywood DISASTERS of 2022
https://youtu.be/PtYtcqSlOcQ