Complaint / review text:
As a white man, I am pissed of watching the hollywood films just include scenes that will please black people (especially black men) and disturb and insult white people (especially white men). Almost in all Hollywood films, they always put standard scenes like:
- black guy being the boss/leader and commanding all the white people around
- black men having relationship, sex with white girls/women (when was the last time somebody saw a black couple in a film)
- black people taking only the good characters like angels
In almost all films (even films like Avengers, Prometheus, Hunger Games etc) these scenes are there.
White people never end a film leaving with a good mood.
It is actually coming to the point that white people should not watch any films at all, not to feel disturbed and insulted (may be, old films only)
Now, the question is WHY? Are hollywood companies getting strict orders from government or so to do so?