America is indeed a divided nation and Hollywood has certainly played a role in this division. Hollywood, as one of the most influential institutions in the world, has a responsibility to reflect the society in which it exists. However, in recent years, it has been accused of promoting a particular political ideology, one that is often referred to as “woke.” …