(This is a copy of a required movie review for Philosophy of Singularity.)
Like most futurist films, Ex Machina tells us more about the current state of humanity than it does the future. Ex Machina, intentionally or unintentionally, was glittered with biases. There were multiple aspects of Ex Machina I found troubling--including (but not limited to): moral, social, and technological issues. Ex Machina is a dystopian view of AI, or rather, a dystopian reflection of humanity.
While optimistic estimates put human-level AI at 2029, the most important thing we can do is reflect upon ourselves and question what we are programming into our creations. Some contemplate developing an ethical code for AI integration and creation, but in all honesty, I’m inclined to believe that the ethical code we need to be evaluating is our own. Our AI will be reflections of ourselves. What do we want that to look like?
For example, Ava, the AI, was played by a white woman. Our moral codes and social structures have a long history with its treatment of women. In Ex Machina, Nathan, the creator of Ava, hired Caleb to determine if Ava passed for human-level intelligence, a Turing test of sorts. Historically, there have been more than a few occasions when men have been in a disproportionate position of deciding when women are “allowed” to the full status of personhood, e.g. legally, intellectually, or physically. Granting the status of personhood in Ex Machina was determined by two men, which could be a reflection of historic and present representations of women.
Though Caleb may have had the best of intentions, his biases were quite evident. It was clear throughout the film he cared for the well-being of Ava, but he didn’t seem to have the same concern for Kyoko. There is an intersectional component that shouldn’t be overlooked. Kyoko, a woman of color, who was specifically limited in her ability to speak, was treated far worse than Ava. She was treated at best as a house servant, and at worst as a sex slave. The message received was, remove a woman’s ability to communicate and you remove her personhood, autonomy, and consent.
When Kyoko attempted to take off her clothes when she and Caleb were alone, Caleb, and likely the majority of the audience, assumed this as a sexual advance. But could there be another explanation? Was she attempting to communicate to Caleb her real identity? Was she attempting to share something intimate and profoundly important of her personhood that was mistaken as something sexual? Maybe. Couldn’t she have simply ripped the skin off her face? Maybe. Like with any victim of sexual abuse and trauma, perhaps Kyoko was acting out of fear. Her sexuality (and Ava’s for that matter) was portrayed in the movie as one of, if not her most powerful asset. Perhaps it was her sexuality that put her in a position which she was brave enough to expose her true self as an AI.
Either way, Caleb showed little to no indication that he was interested in liberating women like Kyoto, but rather interested in his own gratification with Ava. Was Ava only using Caleb for her escape? Or did Kyoko communicate to Ava Caleb’s lack of desire to liberate women like Kyoko? Is that why Ava left Caleb? We don’t know.
In the end, Kyoko died while Ava lived. After Kyoko’s death, Ava literally took the skin off another woman of color’s body and kept it for her own. I’m having a hard time seeing how this isn’t an awful unintentional bias or a direct representation of white women’s perceptions of women of color. It’s truly horrifying.
While some feminists applaud Ava’s escape, the story teaches us that in order for a woman, a white woman, to attain personhood, autonomy, and freedom, it is at the hand of violence, deceit, betrayal, death, and the expense of women of color and men. If this is how patriarchy is to be defeated, I’m not so sure we’ve achieved anything of value. What we have on our hands is a quasi-Animal Farm scenario where one dictatorship is overthrown just to be replaced by another one. Is there really any justice or improvement at that point? If one demographic is held up at the expense of another, we have not evolved and THAT is the real dystopian message of the film. Not that Ava, the most powerful AI created has escaped her cage, but that there is no genuine progression—just the enhancement of technologies that we program with our own biases, flaws, and oppressions. Ava is exactly what we made her to be. If that’s really what we allow patriarchy to do to us, all of us, then we all lose.
The development of AI needs our intersectional experiences and differences, but more importantly it needs our own internal transformations. AI will be an amplified version of ourselves, all that is good and beautiful and all that is harmful and terrible. We must continue to change ourselves, and not just change, but improve through genuine progression if there is any hope for AI or our species. Radical compassion, love, and benevolence are required for progress, not just a race track to the most advanced technologies. Because without those attributes, there is no advanced civilization.
Notes and Citations
 Ray Kurzweil, “Don’t Fear Artificial Intelligence,” TIME, December 19 2014, accessed May 27, 2017, “The median view of AI practitioners today is that we are still several decades from achieving human-level AI. I am more optimistic and put the date at 2029, but either way, we do have time to devise ethical standards.” http://time.com/3641921/dont-fear-artificial-intelligence/
 Wikipedia, The Free Encyclopedia, s.v. “Turing test,” accessed May 27, 2017 https://en.wikipedia.org/wiki/Turing_test
 J.A. Micheline, “Ex Machina: A (White) Feminist Parable of Our Time,” Women Write About Comics, May 21, 2015, accessed May 27, 2017, “But here’s what’s killer about the android/artificial intelligence = woman metaphor, particularly as shown by Ava. It sucks because it implies that we were created by men and for men, when of course, we were not. And yet, it triumphs because it says that even if we allow such a ridiculous premise, even if we entertain the notion that men are constantly building or trying to build us into what we want to be, it doesn’t matter. Ava, patriarchal dream as she is, shrugs off her programming, shrugs off the way that men want to see her and gets her goddamn own. Which is inspiring, of course, because in the end, what it really means is: so can we.”
 Wikipedia, The Free Encyclopedia, s.v. “Animal Farm,” accessed May 27, 2017 https://en.wikipedia.org/wiki/Animal_Farm
 Clare Garvie and Jonathan Frankle, “Facial-Recognition Software Might Have a Racial Bias Problem,” The Atlantic, April 7, 2016, accessed May 27, 2017 https://www.theatlantic.com/technology/archive/2016/04/the-underlying-bias-of-facial-recognition-systems/476991/