Peter Debruge Chief Film CriticThere’s a cruel irony working against Hollywood’s efforts to diversify: For nearly a century, the industry depicted the world as a place dominated by white, straight, able-bodied men.
The movies typically relegated women and people of color to supporting and subservient roles, while excluding (or else vilifying) queer and handicapped characters.
As a result, entire generations have been raised on lopsided and inaccurate representations of our past — that Jesus was white, for example — to the extent that they don’t necessarily believe it when Black actors appear in situations where they played a prominent (off-screen) role.
Read more on variety.com