Matching Items (2)
133894-Thumbnail Image.png
Description
Pandora is a play exploring our relationship with gendered technology through the lens of artificial intelligence. Can women be subjective under patriarchy? Do robots who look like women have subjectivity? Hoping to create a better version of ourselves, The Engineer must navigate the loss of her creation, and Pandora must

Pandora is a play exploring our relationship with gendered technology through the lens of artificial intelligence. Can women be subjective under patriarchy? Do robots who look like women have subjectivity? Hoping to create a better version of ourselves, The Engineer must navigate the loss of her creation, and Pandora must navigate their new world. The original premiere run was March 27-28, 2018, original cast: Caitlin Andelora, Rikki Tremblay, and Michael Tristano Jr.
ContributorsToye, Abigail Elizabeth (Author) / Linde, Jennifer (Thesis director) / Abele, Kelsey (Committee member) / Department of Information Systems (Contributor) / Economics Program in CLAS (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
148032-Thumbnail Image.png
Description

Artificial Intelligence’s facial recognition programs are inherently racially biased. The programs are not necessarily created with the intent to disproportionately impact marginalized communities, but through their data mining process of learning, they can become biased as the data they use may train them to think in a biased manner. Biased

Artificial Intelligence’s facial recognition programs are inherently racially biased. The programs are not necessarily created with the intent to disproportionately impact marginalized communities, but through their data mining process of learning, they can become biased as the data they use may train them to think in a biased manner. Biased data is difficult to spot as the programming field is homogeneous and this issue reflects underlying societal biases. Facial recognition programs do not identify minorities at the same rate as their Caucasian counterparts leading to false positives in identifications and an increase of run-ins with the law. AI does not have the ability to role-reverse judge as a human does and therefore its use should be limited until a more equitable program is developed and thoroughly tested.

ContributorsGurtler, Charles William (Author) / Iheduru, Okechukwu (Thesis director) / Fette, Donald (Committee member) / Economics Program in CLAS (Contributor) / School of Politics and Global Studies (Contributor) / Barrett, The Honors College (Contributor)
Created2021-05