A sea change is happening in Hollywood’s depiction of women. It may be slow, but it is steady and gaining momentum. For years, women both in front of and behind the camera have been marginalized, stereotyped, and underrepresented, but 2017 marked a turning point.