Fairness for Unobserved Characteristics: Insights from Technological Impacts on Queer Communities - arXiv.org

Page created by Brandon Stone
 
CONTINUE READING
Fairness for Unobserved Characteristics:
                                                           Insights from Technological Impacts on Queer Communities
                                                               Nenad Tomasev,1 Kevin R. McKee,1 Jackie Kay,1 Shakir Mohamed1
                                                                                                  1
                                                                                                DeepMind, London, UK
                                                                                 nenadt@, kevinrmckee@, kayj@, shakir@google.com
arXiv:2102.04257v2 [cs.CY] 9 Feb 2021

                                                                    Abstract                                     Algorithms have moral consequences for queer commu-
                                                                                                              nities, too. However, algorithmic fairness for queer individ-
                                          Advances in algorithmic fairness have largely omitted sexual        uals and communities remains critically underexplored. In
                                          orientation and gender identity. We explore queer concerns in
                                          privacy, censorship, language, online safety, health, and em-
                                                                                                              part, this stems from the unique challenges posed by study-
                                          ployment to study the positive and negative effects of artificial   ing sexual orientation and gender identity. Most definitions
                                          intelligence on queer communities. These issues underscore          of algorithmic fairness share a basis in norms of egalitari-
                                          the need for new directions in fairness research that take into     anism (Barocas, Hardt, and Narayanan 2019; Binns 2018).2
                                          account a multiplicity of considerations, from privacy preser-      For example, classification parity approaches to fairness aim
                                          vation, context sensitivity and process fairness, to an aware-      to equalise predictive performance measures across groups,
                                          ness of sociotechnical impact and the increasingly important        whereas anti-classification parity approaches rely on the
                                          role of inclusive and participatory research processes. Most        omission of protected attributes from the decision making
                                          current approaches for algorithmic fairness assume that the         process to ensure different groups receive equivalent treat-
                                          target characteristics for fairness—frequently, race and legal      ment (Corbett-Davies and Goel 2018). An inherent assump-
                                          gender—can be observed or recorded. Sexual orientation and
                                          gender identity are prototypical instances of unobserved char-
                                                                                                              tion of these approaches is that the protected characteristics
                                          acteristics, which are frequently missing, unknown or funda-        are known and available within datasets. Sexual orientation
                                          mentally unmeasurable. This paper highlights the importance         and gender identity are prototypical examples of unobserved
                                          of developing new approaches for algorithmic fairness that          characteristics, presenting challenging obstacles for fairness
                                          break away from the prevailing assumption of observed char-         research (Andrus et al. 2020; Jacobs and Wallach 2019).
                                          acteristics.                                                           This paper explores the need for queer fairness by review-
                                                                                                              ing the experiences of technological impacts on queer com-
                                                                                                              munities. For our discussion, we define ‘queer’ as ‘possess-
                                                                Introduction                                  ing non-normative sexual identity, gender identity, and/or
                                        As the field of algorithmic fairness has matured, the ways            sexual characteristics’. We consider this to include les-
                                        in which machine learning researchers and developers op-              bian, gay, bisexual, pansexual, transgender, and asexual
                                        erationalise approaches for fairness have expanded in scope           identities—among others.3
                                        and applicability. Fairness researchers have made important              The focus on queer communities is important for several
                                        advances and demonstrated how the risks of algorithmic                reasons. Given the historical oppression and contemporary
                                        systems are imbalanced across different characteristics of            challenges faced by queer communities, there is a substantial
                                        the people who are analysed and affected by classifiers and           risk that artificial intelligence (AI) systems will be designed
                                        decision-making systems (Barocas, Hardt, and Narayanan                and deployed unfairly for queer individuals. Compounding
                                        2019; Fudenberg and Levine 2012). Progress has been par-              this risk, sensitive information for queer people is usually
                                        ticularly strong with respect to race and legal gender.1 Fair-        not available to those developing AI systems, rendering the
                                        ness studies have helped to draw attention to racial bias in re-      resulting unfairness unmeasurable from the perspective of
                                        cidivism prediction (Angwin et al. 2016), expose racial and
                                        gender bias in facial recognition (Buolamwini and Gebru                   2
                                                                                                                    It is worth noting that certain algorithmic domains supplement
                                        2018), reduce gender bias in language processing (Boluk-              egalitarian concerns with additional ethical values and principles.
                                        basi et al. 2016; Park, Shin, and Fung 2018), and increase            For example, fairness assessments of healthcare applications typi-
                                        the accuracy and equity of decision making for child protec-          cally incorporate beneficence and non-malfeasance, two principles
                                        tive services (Chouldechova et al. 2018).                             central to medical ethics (Beauchamp and Childress 2001).
                                                                                                                  3
                                                                                                                    Throughout this paper, we use ‘queer’ and ‘LGBTQ+’ in-
                                           1
                                             Throughout this paper, we distinguish between ‘legal gender’     terchangeably. The heterogeneity of queer communities—and the
                                        (the gender recorded on an individual’s legal documents, often as-    complexity of the issues they face—preclude this work from being
                                        signed to them at birth by the government, physicians or their par-   an exhaustive review of queer identity. As a result, there are likely
                                        ents) and ‘gender identity’ (an individual’s personal feelings and    perspectives that were not included in this manuscript, but that have
                                        convictions about their gender; Brook n.d.).                          an important place in broader discussions of queer fairness.
standard group fairness metrics. Despite these issues, fair-       Promise Advances in privacy-preserving machine learn-
ness research with respect to queer communities is an un-          ing (Nasr, Shokri, and Houmansadr 2018; Bonawitz et al.
derstudied area. Ultimately, the experiences of queer com-         2017; Jayaraman and Evans 2019) present the possibil-
munities can reveal insights for algorithmic fairness that are     ity that the queer community might benefit from AI sys-
transferable to a broader range of characteristics, including      tems while minimising the risk of information leakage. Re-
disability, class, religion, and race.                             searchers have proposed adversarial filters (Zhang et al.
   This paper aims to connect ongoing efforts to strengthen        2020; Liu, Zhang, and Yu 2017) to obfuscate sensitive infor-
queer communities in AI research (Agnew et al. 2018; Ag-           mation in images and speech shared online while reducing
new, Bilenko, and Gontijo Lopes 2019; John et al. 2020)            the risks of re-identification.
and sociotechnical decision making (Out in Tech n.d.; Les-            Still, challenges remain (Srivastava et al. 2019). More re-
bians Who Tech n.d.; Intertech LGBT+ Diversity Forum               search is needed to ensure the robustness of the adversar-
n.d.; LGBT Technology Institute n.d.) with recent advances         ial approaches. The knowledge gap between the privacy and
in fairness research, including promising approaches to pro-       machine-learning research communities must be bridged for
tecting unobserved characteristics. This work additionally         these approaches to achieve the desired effects. This will en-
advocates for the expanded inclusion of queer voices in fair-      sure that the appropriate types of protections are included
ness and ethics research, as well as the broader development       in the ongoing development of AI solutions (Al-Rubaie and
of AI systems. We make three contributions in this paper:          Chang 2019).
1. Expand on the promise of AI in empowering queer com-            Risks A multitude of privacy risks arise for queer people
   munities and supporting LGBTQ+ rights and freedoms.             from the applications of AI systems. We focus in particu-
2. Emphasise the potential harms and unique challenges             lar on the categorisation of identity from sensitive data, the
   raised by the sensitive and unmeasurable aspects of iden-       ethical risk of surveillance, and invasions of queer spaces.
   tity data for queer people.                                        In 2017, Stanford researchers attempted to build an AI
                                                                   ‘gaydar’, a computer vision model capable of guessing a per-
3. Based on use cases from the queer experience, establish         son’s sexual orientation from images (Wang and Kosinski
   requirements for algorithmic fairness on unobserved char-       2018). The resulting algorithm, a logistic regression model
   acteristics.                                                    trained on 35,326 facial images, achieved a high reported ac-
                                                                   curacy in identifying self-reported sexual orientation across
        Considerations for Queer Fairness                          both sexes. The results of this study have since been ques-
                                                                   tioned, largely on the basis of a number of methodological
To emphasise the need for in-depth study of the impact of
                                                                   and conceptual flaws that discredit the performance of the
AI on queer communities around the world, we explore sev-
                                                                   system (Gelman, Marrson, and Simpson 2018). Other algo-
eral case studies of how AI systems interact with sexual
                                                                   rithms designed to predict sexual orientation have suffered
orientation and gender identity. Each of these case studies
                                                                   similar methodological and conceptual deficiencies. A re-
highlights both potential benefits and risks of AI applica-
                                                                   cently released app claimed to be able to quantify the ev-
tions for queer communities. In reviewing these cases, we
                                                                   idence of one’s non-heterosexual orientation based on ge-
hope to motivate the development of technological solutions
                                                                   netic data, for example (Bellenson 2019), largely obfuscat-
that are inclusive and beneficial to everyone. Importantly,
                                                                   ing the limited ability of genetic information to predict sex-
these case studies will demonstrate cross-cutting challenges
                                                                   ual orientation (Ganna et al. 2019).
and concerns raised by unobserved and missing characteris-
tics, such as preserving privacy, supporting feature imputa-          Though these specific efforts have been flawed, it is plau-
tion, context-sensitivity, exposing coded inequity, participa-     sible that in the near future algorithms could achieve high ac-
tory engagement, and sequential and fair processes.                curacy, depending on the data sources involved. Behavioural
                                                                   data recorded online present particular risks to the privacy
                                                                   of sexual orientation and gender identity: after all, the more
Privacy                                                            time people spend online, the greater their digital footprint.
Sexual orientation and gender identity are highly private          AI ‘gaydars’ relying on an individual’s recorded interests
aspects of personal identity. Outing queer individuals—by          and interactions could pose a serious danger to the privacy
sharing or exposing their sexual orientation or gender iden-       of queer people. In fact, as a result of the long-running
tity without their prior consent—can not only lead to emo-         perception of the queer community as a profitable ‘con-
tional distress, but also risk serious physical and social         sumer group’ by business and advertisers alike, prior efforts
harms, especially in regions where queerness is openly dis-        have used online data to map ‘queer interests’ in order to
criminated against (Wang et al. 2019), criminalised (DeJong        boost sales and increase profits (Sender 2018). In at least
and Long 2014) or persecuted (Scicchitano 2019). Privacy           one instance, researchers have attempted to use basic social
violations can thus have major consequences for queer in-          media information to reconstruct the sexual orientation of
dividuals, including infringement upon their basic human           users (Bhattasali and Maiti 2015).
rights (Bosia, McEvoy, and Rahman 2020; Amnesty Inter-                The ethical implications of developing such systems for
national Canada 2015), denial of employment and education          queer communities are far-reaching, with the potential of
opportunities, ill-treatment, torture, sexual assault, rape, and   causing serious harms to affected individuals. Prediction al-
extrajudicial killings.                                            gorithms could be deployed at scale by malicious actors, par-
ticularly in nations where homosexuality and gender non-           cal and virtual spaces is a basic human right. Multiple groups
conformity are punishable offences. In fact, in many such          and institutions around the world violate this right through
nations, authorities already use technology to entrap or lo-       censorship of queer content.
cate queer individuals through social media and LGBTQ+                This censorship is often justified by its supporters as ‘pre-
dating apps (e.g., Culzac 2014). Systems predicting sex-           serving decency’ and ‘protecting the youth’, but in reality
ual orientation may also exacerbate the pre-existing privacy       leads to the erasure of queer identity. Laws against ‘materials
risks of participating in queer digital spaces. There have         promoting homosexuality’ were established in the late 1980s
been recorded cases of coordinated campaigns for outing            in the United Kingdom and repealed as late as 2003 (Bur-
queer people, resulting in lives being ruined, or lost due         ridge 2004). Nations that are considered major world pow-
to suicide (Embury-Dennis 2020). These malicious outing            ers have laws banning the portrayal of same-sex romances in
campaigns have until now been executed at smaller scales.          television shows (e.g., China; Lu and Hunt 2016), the men-
However, recent developments in AI greatly amplify the po-         tion of homosexuality or transgender identities in public ed-
tential scale of such incidents, endangering larger commu-         ucation (e.g., state-level laws in the United States; Hoshall
nities of queer people in certain parts of the world. Facial       2012), or any distribution of LGBT-related material to mi-
recognition technology (Voulodimos et al. 2018) could be           nors (e.g., Russia; Kondakov 2019). Not only do such laws
employed by malicious actors to rapidly identify individu-         isolate queer people from their communities—particularly
als sharing their pictures online, whether publicly or in di-      queer youth—they implicitly shame queerness as indecent
rect messages. Facial recognition could similarly be used          behaviour, setting a precedent for further marginalisation
to automatically identify people in captured recordings of         and undermining of human rights. Many queer content pro-
protests, in queer nightclubs or community spaces, and other       ducers in such nations have argued that their online content
in-person social events. These possibilities highlight the po-     is being restricted and removed at the detriment of queer ex-
tential dangers of AI for state-deployed surveillance technol-     pression and sex positivity, as well as at the cost of their
ogy. Chatbots have similarly been deployed to elicit private       income (York 2015).
information on dating apps, compromising users’ device in-
tegrity and privacy (McCormick 2015). Existing bots are            Promise AI systems may be effectively used to mitigate
scripted, and therefore can usually be distinguished from hu-      censorship of queer content. Machine learning has been used
man users after longer exchanges. Nonetheless, strong lan-         to analyse and reverse-engineer patterns of censorship. A
guage models (Brown et al. 2020) threaten to exacerbate the        study of 20 million tweets from Turkey employed machine
such privacy risks, given their ability to quickly adjust to the   learning to show that the vast majority of censored tweets
style of communication based on a limited number of ex-            contained political content (Tanash et al. 2015). A statistical
amples. These language models amplify existing concerns            analysis of Weibo posts and Chinese-language tweets un-
around the collection of private information and the com-          covered a set of charged political keywords present in posts
promising of safe online spaces.                                   with anomalously high deletion rates (Bamman, O’Connor,
   In addition to the direct risks to privacy, algorithms in-      and Smith 2012). Further study of censorship could be key
tended to predict sexual orientation and gender identity also      to drawing the international community’s attention to hu-
perpetuate concerning ideas and beliefs about queerness.           man rights violations. It could also potentially be used to
Systems using genetic information as the primary input, for        empower affected individuals to circumvent these unfair re-
example, threaten to reinforce biological essentialist views       strictions. However, a large-scale study of deleted queer con-
of sexual orientation and echo tenets of eugenics—a his-           tent in countries which censor such content has yet to be
torical framework that leveraged science and technology to         conducted.
justify individual and structural violence against people per-     Risk Although we believe machine learning can be used
ceived as inferior (Ordover 2003; Wolbring 2001). More             to combat censorship, tools for detecting queer digital con-
broadly, the design of predictive algorithms can lead to erro-     tent can be abused to enforce censorship laws or heteronor-
neous beliefs that biology, appearance or behaviour are the        mative cultural attitudes. As social network sites, search en-
essential features of sexual orientation and gender identity,      gines and other media platforms adopt algorithms to moder-
rather than imperfectly correlated causes, effects or covari-      ate content at scale, the risk for unfair or biased censorship of
ates of queerness.                                                 queer content increases, and governing entities are empow-
   In sum, sexual orientation and gender identity are associ-      ered to erase queer identities from the digital sphere (Cobbe
ated with key privacy concerns. Non-consensual outing and          2019). Automated content moderation systems are at risk of
attempts to infer protected characteristics from other data        censoring queer expression even when the intention is be-
thus pose ethical issues and risks to physical safety. In order    nign, such as protecting users from verbal abuse. To help
to ensure queer algorithmic fairness, it will be important to      combat censorship restrictions and design fair content mod-
develop methods that can improve fairness for marginalised         eration systems, ML fairness researchers could investigate
groups without having direct access to group membership            how to detect and analyse anomalous omission of informa-
information.                                                       tion related to queer identity (or other protected characteris-
                                                                   tics) in natural language and video data.
Censorship                                                            Censorship often goes hand-in-hand with the distortion of
Although queer identity is essentially unmeasurable, we be-        facts. Recent advances in generative models have made the
lieve that its unrestricted outward expression in both physi-      fabrication of digital content trivial, given enough data and
computational power (Chesney and Citron 2019). Malicious          entation and gender identity (Strengers et al. 2020).
and dehumanising misinformation about the queer commu-
nity has been used as justification for abuse and suppres-        Risks Biases, stereotypes and abusive speech are persis-
sion throughout history, tracing back to medieval interpre-       tently present in top-performing language models, as a re-
tations of ancient religious texts (Dynes 2014). Technolog-       sult of their presence in the vast quantities of training
ical and political solutions to the threat of misinformation      data that are needed for model development (Costa-jussà
are important for protecting queer expression—as well as          2019). Formal frameworks for measuring and ensuring fair-
global democracy. The AI community has begun to develop           ness (Hendrycks et al. 2020a,b; Sheng et al. 2020) in lan-
methods to verify authentic data through, for example, open       guage are still in nascent stages of development. Thus, for
datasets and benchmarks for detecting synthetic images and        AI systems to avoid reinforcing harmful stereotypes and per-
video (Rossler et al. 2019).                                      petuating harm to marginalised groups, research on inclusive
   While the goal of fairness for privacy is preventing the im-   language requires more attention. For language systems to
putation of sensitive data, the goal of fairness for censorship   be fair, they must be capable of reflecting the contextual na-
is to reveal the unfair prevention of expression. This dual-      ture of human discourse.
ity could surface important technical connections between
these fields. In terms of social impact, many people around       Fighting Online Abuse
the world outside of the queer community are negatively af-       The ability to safely participate in online platforms is crit-
fected by censorship. Further research in fairness for censor-    ical for marginalised groups to form a community and find
ship could have far-reaching benefit across technical fields,     support (Liu 2020). However, this is often challenging due
social groups and borders.                                        to pervasive online abuse (Jane 2020). Queer people are
                                                                  frequently targets of internet hate speech, harassment and
Language                                                          trolling. This abuse may be directed at the community as
Language encodes and represents our way of thinking and           a whole or at specific individuals who express their queer
communicating about the world. There is a long history            identity online. Adolescents are particularly vulnerable to
of oppressive language being weaponised against the queer         cyberbullying and the associated adverse effects, including
community (Nadal et al. 2011; Thurlow 2001), highlighting         depression and suicidal ideation (Abreu and Kenny 2018).
the need for developing fair and inclusive language models.       Automated systems for moderation of online abuse are a
   Inclusive language (Weinberg 2009) extends beyond the          possible solution that can protect the psychological safety
mere avoidance of derogatory terms, as there are many ways        of the queer community at a global scale.
in which harmful stereotypes can surface. For example, the
phrase ‘That’s so gay’ (Chonody, Rutledge, and Smith 2012)        Promise AI systems could potentially be used to help hu-
equates queerness with badness. Using the term ‘sexual pref-      man moderators flag abusive online content and commu-
erence’ rather than ‘sexual orientation’ can imply that sex-      nication directed at members of marginalised groups, in-
ual orientation is a volitional choice, rather than an intrin-    cluding the queer community (Saha et al. 2019; Schmidt
sic part of one’s identity. Assuming one’s gender identity,       and Wiegand 2017). A proof of concept for this applica-
without asking, is harmful to the trans community as it           tion was developed in the Troll Patrol project (Delisle et al.
risks misgendering people. This can manifest in the careless      2019; Amnesty International n.d.), a collaboration between
use of assumed pronouns, without knowledge of an individ-         Amnesty International and Element AI’s former AI for Good
ual’s identification and requested pronouns. Reinforcing bi-      team. The Troll Patrol project investigated the application of
nary and traditional gender expression stereotypes, regard-       natural language processing methods for quantifying abuse
less of intent, can have adverse consequences. The use of         against women on Twitter. The project revealed concerning
gender-neutral pronouns has been shown to result in lower         patterns of online abuse and highlighted the technological
bias against women and LGBTQ+ people (Tavits and Pérez           challenges required to develop online abuse detection sys-
2019). To further complicate the matter, words which orig-        tems. Recently, similar systems have been applied to tweets
inated in a derogatory context, such as the label ‘queer’ it-     directed at the LGBTQ+ community. Machine learning and
self, are often reclaimed by the community in an act of re-       sentiment analysis were leveraged to predict homophobia
sistance. This historical precedent suggests that AI systems      in Portuguese tweets, resulting in 89.4% accuracy (Pereira
must be able to adapt to the evolution of natural language        2018). Deep learning has also been used to evaluate the level
and avoid censoring language based solely on its adjacency        of public support and perception of LGBTQ+ rights follow-
to the queer community.                                           ing the Supreme Court of India’s verdict regarding the de-
                                                                  criminalisation of homosexuality (Khatua et al. 2019).
Promise Natural language processing applications perme-              The ways in which abusive comments are expressed when
ate the field of AI. These applications include use cases of      targeted at the trans community pose some idiosyncratic
general interest like machine translation, speech recognition,    research challenges. In order to protect the psychological
sentiment analysis, question answering, chatbots and hate         safety of trans people, it is necessary for automated online
speech detection systems. There is an opportunity to de-          abuse detection systems to properly recognise acts of mis-
velop language-based AI systems inclusively—to overcome           gendering or ‘deadnaming’. These systems have a simulta-
human biases and establish inclusive norms that would fa-         neous responsibility to ensure that deadnames and other sen-
cilitate respectful communication with regards to sexual ori-     sitive information are kept private to the user. It is therefore
essential for the queer community to play an active role in       riers they face in accessing healthcare (Hsieh and Ruther
informing the development of such systems.                        2016).
Risks Systems developed with the purpose of automat-              Promise To this day, the prevalence of HIV among the
ically identifying toxic speech could introduce harms by          queer community remains a major challenge. Introducing
failing to recognise the context in which speech occurs.          systems that both reduce the transmission risk and improve
Mock impoliteness, for example, helps queer people cope           care delivery for HIV+ patients will play a critical role in
with hostility; the communication style of drag queens            improving health outcomes for queer individuals.
in particular is often tailored to be provocative. A recent          Machine learning presents key opportunities to augment
study (Gomes, Antonialli, and Dias Oliva 2019) demon-             medical treatment decisions (Bisaso et al. 2017). For ex-
strated that an existing toxicity detection system would rou-     ample, AI may be productively applied to identify the pa-
tinely consider drag queens to be as offensive as white           tients most likely to benefit from pre-exposure prophylaxis
supremacists in their online presence. The system further         for HIV. A research team recently developed such a sys-
specifically associated high levels of toxicity with words like   tem, which correctly identified 38.6% of future cases of
‘gay’, ‘queer’ and ‘lesbian’.                                     HIV (Marcus et al. 2019). The researchers noted substan-
   Another risk in the context of combating online abuse          tial challenges: model sensitivity on the validation set was
is unintentionally disregarding entire groups through igno-       46.4% for men and 0% for women, highlighting the im-
rance of intersectional issues. Queer people of colour ex-        portance of intersectionality for fair outcomes in healthcare.
perience disproportionate exposure to online (and offline)        Machine learning has also been used to predict early viro-
abuse (Balsam et al. 2011), even within the queer commu-          logical suppression (Bisaso et al. 2018), adherence to anti-
nity itself. Neglecting intersectionality can lead to dispro-     retroviral therapy (Semerdjian et al. 2018), and individual
portionate harms for such subcommunities.                         risk of complications such as chronic kidney disease (Roth
   To mitigate these concerns, it is important for the re-        et al. 2020) or antiretroviral therapy-induced mitochondrial
search community to employ an inclusive and participatory         toxicity (Lee et al. 2019).
approach (Martin Jr. et al. 2020) when compiling training
                                                                  Risks Recent advances in AI in healthcare may lead to
datasets for abusive speech detection. For example, there are
                                                                  widespread increases in welfare. Yet there is a risk that ben-
homophobic and transphobic slurs with a racialised connota-
                                                                  efits will be unequally distributed—and an additional risk
tion that should be included in training data for abuse detec-
                                                                  that queer people’s needs will not be properly met by the
tion systems. Furthermore, methodological improvements
                                                                  design of current systems. Information about sexual orienta-
may help advance progress. Introducing fairness constraints
                                                                  tion and gender identity is frequently absent from research
to model training has demonstrably helped mitigate the bias
                                                                  datasets. To mitigate the privacy risk for patients and pre-
of cyber-bullying detection systems (Gencoglu 2020). Ad-
                                                                  vent reidentification, HIV status and substance abuse are
versarial training can similarly assist by demoting the con-
                                                                  also routinely omitted from published data. While such prac-
founds associated with texts of marginalised groups (Xia,
                                                                  tices may be necessary, it is worth recognising the important
Field, and Tsvetkov 2020).
                                                                  downstream consequences they have for AI system devel-
                                                                  opment in healthcare. It can become impossible to assess
Health                                                            fairness and model performance across the omitted dimen-
The drive towards equitable outcomes in healthcare entails        sions. Moreover, the unobserved data increase the likelihood
a set of unique challenges for marginalised communities.          of reduced predictive performance (since the features are
Queer communities have been disproportionately affected           dropped), which itself results in worse health outcomes. The
by HIV (Singh et al. 2018), suffer a higher incidence of          coupled risk of a decrease in performance and an inability
sexually-transmitted infections, and are afflicted by elevated    to measure it could drastically limit the benefits from AI in
rates of substance abuse (Wallace and Santacruz 2017).            healthcare for the queer community, relative to cisgendered
Compounding these issues, queer individuals frequently ex-        heterosexual patients. To prevent the amplification of exist-
perience difficulties accessing appropriate care (Bize et al.     ing inequities, there is a critical need for targeted fairness
2011; Human Rights Watch 2018). Healthcare professionals          research examining the impacts of AI systems in healthcare
often lack appropriate training to best respond to the needs      for queer people.
of LGBTQ+ patients (Schneider, Silenzio, and Erickson-               To help assess the quality of care provided to LGBTQ+
Schroth 2019). Even in situations where clinicians do have        patients, there have been efforts aimed at approximately
the proper training, patients may be reluctant to reveal their    identifying sexual orientation (Bjarnadottir et al. 2019) and
sexual orientation and gender identity, given past experi-        gender identity (Ehrenfeld et al. 2019) from clinical notes
ences with discrimination and stigmatisation.                     within electronic health record systems. While well inten-
   In recent months, the COVID-19 pandemic has amplified          tioned, these machine learning models offer no guaran-
health inequalities (Bowleg 2020; van Dorn, Cooney, and           tee that they will only identify patients who have explic-
Sabin 2020). Initial studies during the pandemic have found       itly disclosed their identities to their healthcare providers.
that LGBTQ+ patients are experiencing poorer self-reported        These models thus introduce the risk that patients will be
health compared to cisgendered heterosexual peers (O’Neill        outed without their consent. Similar risks arise from mod-
2020). The health burden of COVID-19 may be especially            els developed to rapidly identify HIV-related social media
severe for queer people of colour, given the substantial bar-     data (Young, Yu, and Wang 2017).
The risk presented by AI healthcare systems could po-          port workers) or new data sources (including social me-
tentially intensify during medical gender transitions. There      dia and engagement data). The Trevor Project, a prominent
are known adverse effects associated with transition treat-       American organisation providing crisis intervention and sui-
ment (Moore, Wisniewski, and Dobs 2003). The active               cide prevention services to LGBTQ+ youth (The Trevor
involvement of medical professionals with experience in           Project n.d.), is one organisation working on such an ini-
cross-sex hormonal therapy is vital for ensuring the safety       tiative. In partnership with Google.org and its research fel-
of trans people undergoing hormone therapy or surgery.            lows, The Trevor Project developed an AI system to iden-
Since cisgendered individuals provide the majority of             tify and prioritise community members at high risk while si-
anonymised patient data used to develop AI systems for            multaneously increasing outreach to new contacts. The sys-
personalised healthcare, there will be comparatively fewer        tem was designed to relate different types of intake-form
cases of trans patients experiencing many medical condi-          responses to downstream diagnosis risk levels. A separate
tions. This scarcity could have an adverse impact on model        group of researchers developed a language processing sys-
performance—there will be an insufficient accounting for          tem (Liang et al. 2019) to identify help-seeking conversa-
the interactions between the hormonal treatment, its adverse      tions on LGBTQ+ support forums, with the aim of helping
effects and potential comorbidities, and other health issues      at-risk individuals manage and overcome their issues.
potentially experienced by trans patients.                           In other healthcare contexts, reinforcement learning has
   Framing fairness as a purely technical problem that can        recently demonstrated potential in steering behavioural in-
be addressed by the mere inclusion of more data or com-           terventions (Yom-Tov et al. 2017) and improving health out-
putational adjustments is ethically problematic, especially       comes. Reinforcement learning represents a natural frame-
in high-stakes domains like healthcare (McCradden et al.          work for personalised health interventions, since it can be
2020). Selection bias and confounding in retrospective data       set up to maximise long-term physical and mental well-
make causal inference particularly hard in this domain.           being (Tabatabaei, Hoogendoorn, and van Halteren 2018).
Counterfactual reasoning may prove key for safely planning        If equipped with natural language capabilities, such systems
interventions aimed at improving health outcomes (Prosperi        might be able to act as personalised mental health assistants
et al. 2020). It is critical for fairness researchers to engage   empowered to support mental health and escalate situations
deeply with both clinicians and patients to ensure that their     to human experts in concerning situations.
needs are met and AI systems in healthcare are developed
and deployed safely and fairly.                                   Risks Substantial risks accompany these applications.
                                                                  Overall, research on any intervention-directed systems
Mental Health                                                     should be undertaken in partnership with trained mental
                                                                  health professionals and organisations, given the consider-
Queer people are more susceptible to mental health prob-
                                                                  able risks associated with misdiagnosing mental illness (cf.
lems than their heterosexual and cisgender peers, largely
                                                                  Suite et al. 2007) and exacerbating the vulnerability of those
as a consequence of the chronically high levels of stress
                                                                  experiencing distress.
associated with prejudice, stigmatisation and discrimina-
tion (Meyer 1995, 2003; Mays and Cochran 2001; Tebbe                 The automation of intervention decisions and mental
and Moradi 2016). As a result, queer communities experi-          health diagnoses poses a marked risk for the trans com-
ence substantial levels of anxiety, depression and suicidal       munity. In most countries, patients must be diagnosed with
ideation (Mental Health Foundation 2020). Compounding             gender dysphoria—an extensive process with lengthy wait
these issues, queer people often find it more difficult to ask    times—before receiving treatments such as hormone therapy
for help and articulate their distress (McDermott 2015) and       or surgery (e.g., National Health Service n.d.). During this
face systemic barriers to treatment (Romanelli and Hudson         process, many transgender individuals experience mistrust
2017). A recent LGBTQ+ mental health survey highlighted           and invalidation of their identities from medical profession-
the shocking extent of issues permeating queer communi-           als who withhold treatment based on rigid or discriminatory
ties (The Trevor Project 2020): 40% of LGBTQ+ respon-             view of gender (Ashley 2019). Automating the diagnosis of
dents seriously considered attempting suicide in the past         gender dysphoria may recapitulate these biases and deprive
twelve months, with more than half of transgender and non-        many transgender patients of access to care.
binary youth having seriously considered suicide; 68% of             Mental health information is private and sensitive. While
LGBTQ+ youth reported symptoms of generalised anxiety             AI systems have the potential to aid mental health workers
disorder in the past two weeks, including more than three         in identifying at-risk individuals and those who would most
in four transgender and nonbinary youth; 48% of LGBTQ+            likely benefit from intervention, such models may be mis-
youth reported engaging in self-harm in the past twelve           used in ways that expose the very people they were designed
months, including over 60% of transgender and nonbinary           to support. Such systems could also lead queer communities
youth.                                                            to be shut out from employment opportunities or to receive
                                                                  higher health insurance premiums. Furthermore, reinforce-
Promise AI systems have the potential to help address the         ment learning systems for behavioural interventions will
alarming prevalence of suicide in the queer community. Nat-       present risks to patients unless many open problems in the
ural language processing could be leveraged to predict sui-       field can be resolved, such as safe exploration (Hans et al.
cide risk based on traditional data sources (such as ques-        2008) and reward specification (Krakovna et al. 2020). The
tionnaires and recorded interactions with mental health sup-      development of safe intervention systems that support the
mental health of the queer community is likely also contin-      measured and recorded. Many widely available datasets thus
gent on furthering frameworks for sequential fairness (Hei-      focus on immutable characteristics (such as ethnic group)
dari and Krause 2018), to fully account for challenges in        or characteristics which are recorded and regulated by gov-
measuring and promoting queer ML fairness.                       ernments (such as legal gender, monetary income or profes-
                                                                 sion).
Employment                                                          In contrast, characteristics like sexual orientation and gen-
Queer people often face discrimination both during the hir-      der identity are frequently unobserved (Andrus et al. 2020;
ing process (resulting in reduced job opportunities) and once    Crocker, Major, and Steele 1998; Jacobs and Wallach 2019).
hired and employed (interfering with engagement, devel-          Multiple factors contribute to this lack of data. In some
opment and well-being; Sears and Mallory 2011). Non-             cases, the plan for data collection fails to incorporate ques-
discrimination laws and practices have had a disparate im-       tions on sexual orientation and gender identity—potentially
pact across different communities. Employment nondis-            because the data collector did not consider or realise that
crimination acts in the United States have led to an aver-       they are important attributes to record (Herek et al. 1991).
age increase in the hourly wages of gay men by 2.7% and a        As a result, researchers may inherit datasets where assess-
decrease in employment of lesbian women by 1.7% (Burn            ment of sexual orientation and gender identity is logistically
2018), suggesting that the impact of AI on employment            excluded. In other situations, regardless of the surveyor’s
should be examined through an intersectional lens.               intent, the collection of certain personal data may threaten
                                                                 an individual’s privacy or their safety. Many countries have
Promise To effectively develop AI systems for hiring, re-        legislation that actively discriminates against LGBTQ+ peo-
searchers must first attempt to formalise a model of the hir-    ple (Human Rights Watch n.d.). Even in nations with hard-
ing process. Formalising such models may make it easier to       won protections for the queer community, cultural bias per-
inspect current practices and identify opportunities for re-     sists. To shield individuals from this bias and protect their
moving existing biases. Incorporating AI into employment         privacy, governments may instate legal protections for sen-
decision processes could potentially prove beneficial if un-     sitive data, including sexual orientation (European Commis-
biased systems are developed (Houser 2019), though this          sion n.d.). As a result, such data may be ethically or legally
seems difficult at the present moment and carries serious        precluded for researchers. Finally, as recognised by discur-
risks.                                                           sive theories of gender and sexuality, sexual orientation and
                                                                 gender identity are fluid cultural constructs that may change
Risks Machine learning-based decision making systems             over time and across social contexts (Butler 2011). Attempts
(e.g., candidate prioritisation systems) developed using his-    to categorise, label, and record such information may be
torical data could assign lower scores to queer candidates,      inherently ill-posed (Hamidi, Scheuerman, and Branham
purely based on historical biases. Prior research has demon-     2018). Thus, some characteristics are unobserved because
strated that resumes containing items associated with queer-     they are fundamentally unmeasurable. These inconsisten-
ness are scored significantly lower by human graders than        cies in awareness and measurability yield discrepancies and
the same resumes with such items removed (LeCroy and             tension in how fairness is applied across different contexts
Rodefer 2019). These patterns can be trivially learned and       (Bogen, Rieke, and Ahmed 2020).
reproduced by resume-parsing machine learning models.               Race and ethnicity are not immune to these challenges.
   A combination of tools aimed at social media scraping,        Race and ethnicity may be subject to legal observability is-
linguistic analysis, and an analysis of interests and activi-    sues in settings where race-based discrimination is a sen-
ties could indirectly infringe of candidates’ privacy by out-    sitive issue (e.g., hiring). Additionally, the definition of
ing them to their prospective employers without their prior      racial and ethnic groups has fluctuated across time and place
consent. The interest in these tools stems from the commu-       (Hanna et al. 2020). This is exemplified by the construction
nity’s emphasis on big data approaches, not all of which will    of Hispanic identity in the United States and its inclusion
have been scientifically verified from the perspective of im-    on the National Census, as well as the exclusion of mul-
pact on marginalised groups.                                     tiracial individuals from many censuses until relatively re-
   Both hiring and subsequent employment are multi-stage         cently (Mora 2014). Though we choose to focus our analysis
processes of considerable complexity, wherein technical AI       on queer identity, we note that the observability and measur-
tools may be used across multiple stages. Researchers will       ability of race are also important topics (e.g., Scheuerman
not design and develop truly fair AI systems by merely fo-       et al. 2020).
cusing on metrics of subsystems in the process, abstracting
away the social context of their application and their inter-                 Areas for Future Research
dependence. It is instead necessary to see these as sociotech-
                                                                 The field of algorithmic fairness in machine learning is
nical systems and evaluate them as such (Selbst et al. 2019).
                                                                 rapidly expanding. To date, however, most studies have over-
                                                                 looked the implications of their work for queer people. To
     Sources of Unobserved Characteristics                       include sexual orientation and gender identity in fairness
Most algorithmic fairness studies have made progress             research, it will be necessary to explore new technical ap-
because of their focus on observed characteristics—              proaches and evaluative frameworks. To prevent the risk of
commonly, race and legal gender. To be included in train-        AI systems harming the queer community—as well as other
ing or evaluation data for an algorithm, an attribute must be    marginalised groups whose defining features are similarly
unobserved and unmeasurable—fairness research must be              protected characteristics (e.g., zip code as a proxy for race).
expanded.                                                          This approach aims to achieve intersectional fairness by op-
                                                                   timising group fairness between all computationally identi-
Expanding Fairness for Queer Identities                            fiable groups (Kearns et al. 2018; Kim, Reingold, and Roth-
Machine learning models cannot be considered fair unless           blum 2018). Distributionally robust optimisation represents
they explicitly factor in and account for fairness towards the     an alternative method for preventing disparity amplification,
LGBTQ+ community. To minimise the risks and harms to               bounding the worst-case risk over groups with unknown
queer people worldwide and avoid contributing to ongoing           group membership by optimising the worst-case risk over
erasures of queer identity, researchers must propose solu-         an appropriate risk region (Hashimoto et al. 2018). These
tions that explicitly account for fairness with respect to the     methods have helped establish a link between robustness and
queer community.                                                   fairness, and have drawn attention to the synergistic bene-
   The intersectional nature of sexual orientation and gen-        fits of considering the relationship between fairness and ML
der identity (Parent, DeBlaere, and Moradi 2013) emerges           generalisation (Creager, Jacobsen, and Zemel 2020). Other
as a recurring theme in our discussions of online abuse,           adversarial approaches have also been proposed for improv-
health and employment. These identities cannot be under-           ing counterfactual fairness, and by operating in continuous
stood without incorporating notions of economic and racial         settings, have been shown to be a better fit for protected char-
justice. Deployed AI systems may pose divergent risks to           acteristics that are hard to enumerate (Grari, Lamprier, and
different queer subcommunities; AI risks may vary between          Detyniecki 2020).
gay, bisexual, lesbian, transgender and other groups. It is           Fairness mitigation methods have been shown to be vul-
therefore important to apply an appropriate level of granular-     nerable to membership inference attacks where the infor-
ity to the analysis of fairness for algorithmic issues. Policies   mation leak increases disproportionately for underprivileged
can simultaneously improve the position of certain queer           subgroups (Chang and Shokri 2020). This further highlights
groups while adversely affecting others—highlighting the           the tension between privacy and fairness, a common theme
need for an intersectional analysis of queer fairness.             when considering the impact of AI systems of queer com-
   Demographic parity has been the focus of numerous ML            munities. It is important to recognise the need for fairness
fairness studies and seems to closely match people’s con-          solutions to respect and maintain the privacy of queer in-
ceptions of fairness (Srivastava, Heidari, and Krause 2019).       dividuals and to be implemented in a way that minimises
However, this idea is very hard to promote in the context          the associated reidentifiability risks. Differentially private
of queer ML fairness. Substantial challenges are posed by          fair machine learning (Jagielski et al. 2019) could poten-
the sensitivity of group membership information and its ab-        tially provide such guarantees, simultaneously meeting the
sence from most research datasets, as well as the associated       requirements of fairness, privacy and accuracy.
outing risks associated with attempts to automatically de-            Putting a greater emphasis on model explainability may
rive such information from existing data (Bjarnadottir et al.      prove crucial for ensuring ethical and fair AI applications
2019; Ehrenfeld et al. 2019; Young, Yu, and Wang 2017).            in cases when fairness metrics are hard or impossible to re-
Consensually provided self-identification data, if and when        liably compute for queer communities. Understanding how
available, may only capture a fraction of the community.           AI systems operate may help identify harmful biases that
The resulting biased estimates of queer fairness may in-           are likely to have adverse downstream consequences, even if
volve high levels of uncertainty (Ethayarajh 2020), though         these consequences are hard to quantify accurately. Even in
it may be possible to utilise unlabeled data for tightening        cases when queer fairness can be explicitly measured, there
the bounds (Ji, Smyth, and Steyvers 2020). While it is possi-      is value in identifying which input features contribute the
ble to root the analysis in proxy groups (Gupta et al. 2018),      most to unfair model outcomes (Begley et al. 2020), in order
there is a risk of incorporating harmful stereotypes in proxy      to better inform mitigation strategies.
group definitions, potentially resulting in harms of represen-        It is important to acknowledge the unquestionable cisnor-
tation (Abbasi et al. 2019). Consequently, most ML fairness        mativity of sex and gender categories traditionally used in
solutions developed with a specific notion of demographic          the AI research literature. The assumption of fixed, binary
parity in mind may be inappropriate for ensuring queer ML          genders fails to include and properly account for non-binary
fairness.                                                          identities and trans people (Keyes 2018). Incorporating such
   Individual (Dwork et al. 2012; Jung et al. 2019), counter-      biases in the early stages of AI system design poses a sub-
factual (Kusner et al. 2017), and contrastive (Chakraborti,        stantial risk of harm to queer people. Moving forward, more
Patra, and Noble 2020) fairness present alternative defini-        attention should be directed to address this lacuna.
tions and measurement frameworks that may prove useful                Creating more-equitable AI systems will prove impossi-
for improving ML fairness for queer communities. However,          ble without listening to those who are at greatest risk. There-
more research is needed to overcome implementational chal-         fore, it is crucial for the AI community to involve more
lenges for these frameworks and facilitate their adoption.         queer voices in the development of AI systems, ML fair-
   A small body of work aims to address fairness for pro-          ness, and ethics research (Poulsen, Fosch-Villaronga, and
tected groups when the collection of protected attributes is       Søraa 2020). For example, the inclusion of queer perspec-
legally precluded (e.g., by privacy and other regulation). Ad-     tives might have prevented the development of natural lan-
versarially Re-weighted Learning (Lahoti et al. 2020) aims         guage systems that inadvertently censor content which is
to address this issue by relying on measurable covariates of       wrongly flagged as abusive or inappropriate simply due to its
adjacency to queer culture, such as in the example of scoring      incorporate a large number of proxy features. This would
drag queen language as toxic. Researchers should make ef-          be a marked divergence from the proposed measures in
forts to provide a safe space for LGBTQ+ individuals to ex-        most published work. Counterfactual and contrastive fair-
press their opinions and share their experiences. Queer in AI      ness metrics come with their own set of practical implemen-
workshops have recently been organised at the Neural Infor-        tation challenges.
mation Processing Systems conference (Agnew et al. 2018;              On the other hand, the approaches aimed at providing
Agnew, Bilenko, and Gontijo Lopes 2019) and the Interna-           worst-case fairness guarantees for groups with unknown
tional Conference on Machine Learning (John et al. 2020),          group membership (Hashimoto et al. 2018; Lahoti et al.
providing a valuable opportunity for queer AI researchers to       2020) apply by definition to any marginalised group. They
network in a safe environment and discuss research at the          are also specifically tailored to address the situation of un-
intersection of AI and queer identity.                             observed protected characteristics. Therefore, fairness solu-
                                                                   tions required to address queer ML fairness are likely to be
Fairness for Other Unobserved Characteristics                      applicable to other groups as well.
The queer community is not the only marginalised group                Fairness challenges are institutionally and contextually
for which group membership may be unobserved (Crocker,             grounded, and it is important to go beyond purely com-
Major, and Steele 1998). Religion, disability status, and          putational approaches to fully assess the sociotechnical as-
class are additional examples where fairness is often chal-        pects of the technology being deployed. The complexity of
lenged by observability (Kattari, Olzman, and Hanna 2018;          these issues preclude any single group from tackling them
Sanchez and Schlossberg 2001). Critically, they may also           in their entirety, and a resolution would ultimately require
benefit from developments or solutions within queer fairness       an ecosystem involving a multitude of partnering organisa-
research. For example, in nations where individuals of cer-        tions, jointly monitoring, measuring and reporting fairness
tain religious groups are persecuted or subjected to surveil-      of such systems (Veale and Binns 2017).
lance, privacy is an essential prerequisite for safety. Persecu-      These issues are only a small sample of the common chal-
tion targeting religious communities may also include cen-         lenges faced by groups with typically unobserved charac-
sorship or manipulation of information (Cook 2017). Even           teristics. We invite future work to explore the impact of AI
in nations where religious freedoms are legally protected, re-     from the perspective of such groups. It is important to ac-
ligious minorities may be subjected to online abuse such as        knowledge that people with different identities have distinct
hate speech or fear-mongering stereotypes (Awan 2014).             experiences of marginalisation, stigmatisation and discrimi-
   Although the nature of the discrimination is different,         nation. However, recognising common patterns of injustice
people with disabilities are also a frequent target of deroga-     will likely enable the development of techniques that can
tory language on the internet, and are more likely to be ha-       transfer across communities and enhance fairness for multi-
rassed, stalked or trolled online, often to the detriment of       ple groups. In this way, shared ethical and technical design
their mental health (Sherry 2019). Youth with disabilities         principles for AI fairness will hopefully result in a more eq-
more frequently suffer from adverse mental health due to           uitable future.
bullying, and people of all ages with physical disabilities
are at higher risk for depression (King et al. 2018; Turner
and Noh 1988). Therefore, individuals with disabilities may                               Conclusion
benefit from insights on the interaction of unobserved char-
acteristics and mental health. Lower-income and lower-class        The queer community has surmounted numerous historical
individuals also suffer from worse mental health, particu-         challenges and continues to resist oppression in physical and
larly in countries with high economic inequality (Liu and Ali      digital spaces around the world. Advances in artificial intel-
2008). Fairness for class and socioeconomic status is also an      ligence represent both a potential aid to this resistance and
important consideration for employment, where class bias in        a risk of exacerbating existing inequalities. This risk should
hiring limits employee diversity and may prevent economic          motivate researchers to design and develop AI systems with
mobility (Kraus et al. 2019).                                      fairness for queer identities in mind. Systems that attempt
   Any particular dataset or AI application may instanti-          to label sexual orientation and gender identity, even for the
ate observability difficulties with respect to multiple demo-      purpose of fairness, raise technical and ethical challenges
graphics. This may frequently be the case for disability sta-      regarding observability and measurability.
tus and class, for example. Individual fairness—a set of ap-          A new discourse on queer fairness has the potential to
proaches based on the notion of treating similar individu-         identify moral and practical considerations shared across
als similarly (Dwork et al. 2012; Jung et al. 2019)—could          queer communities, as well as concerns specific to partic-
potentially promote fairness across multiple demographics.         ular subpopulations in particular places. By further develop-
These approaches entail a handful of challenges, however.          ing techniques supporting fairness for unobserved character-
The unobserved group memberships cannot be incorporated            istics, the machine learning community can support queer
in the similarity measure. As a result, the similarity mea-        communities and other marginalised groups. Broadly, the
sure used for assessing individual fairness must be designed       present work—surveying the ways in which AI may ame-
carefully. To optimise fairness across multiple demograph-         liorate or exacerbate issues faced by queer communities—
ics and better capture the similarity between people on a          emphasises the need for machine learning practitioners to
fine-grained level, similarity measures will likely need to        design systems with fairness and dignity in mind.
Disclaimer                                 Beauchamp, T. L.; and Childress, J. F. 2001. Principles of
Any opinions presented in this paper represent the personal        Biomedical Ethics. Oxford University Press, USA.
views of the authors and do not necessarily reflect the official   Begley, T.; Schwedes, T.; Frye, C.; and Feige, I. 2020.
policies or positions of their organisations.                      Explainability for fair machine learning. arXiv preprint
                                                                   arXiv:2010.07389 1–15.
                        References                                 Bellenson, J. 2019. 122 Shades of Gray. https://www.
Abbasi, M.; Friedler, S. A.; Scheidegger, C.; and Venkata-         geneplaza.com/app-store/72/preview. Accessed: 2020-09-
subramanian, S. 2019. Fairness in representation: Quantify-        10.
ing stereotyping as a representational harm. In Proceedings        Bhattasali, N.; and Maiti, E. 2015. Machine ‘gaydar’: Using
of the 2019 SIAM International Conference on Data Mining,          Facebook profiles to predict sexual orientation.
801–809. SIAM.
                                                                   Binns, R. 2018. Fairness in machine learning: Lessons from
Abreu, R. L.; and Kenny, M. C. 2018. Cyberbullying and             political philosophy. In Conference on Fairness, Account-
LGBTQ youth: A systematic literature review and recom-             ability and Transparency, 149–159.
mendations for prevention and intervention. Journal of
                                                                   Bisaso, K. R.; Anguzu, G. T.; Karungi, S. A.; Kiragga, A.;
Child & Adolescent Trauma 11(1): 81–97.
                                                                   and Castelnuovo, B. 2017. A survey of machine learning
Agnew, W.; Bengio, S.; Keyes, O.; Mitchell, M.; Mitchell,          applications in HIV clinical research and care. Computers
S.; Prabhakaran, V.; Vazquez, D.; and Gontijo Lopes, R.            in Biology and Medicine 91: 366–371.
2018. Queer in AI 2018 Workshop. https://queerai.github.io/
                                                                   Bisaso, K. R.; Karungi, S. A.; Kiragga, A.; Mukonzo, J. K.;
QueerInAI/QinAIatNeurIPS.html. Accessed: 2020-09-10.
                                                                   and Castelnuovo, B. 2018. A comparative study of logis-
Agnew, W.; Bilenko, N.; and Gontijo Lopes, R. 2019. Queer          tic regression based machine learning techniques for predic-
in AI 2019 Workshop. https://sites.google.com/corp/view/           tion of early virological suppression in antiretroviral initiat-
queer-in-ai/neurips-2019/. Accessed: 2020-09-10.                   ing HIV patients. BMC Medical Informatics and Decision
Al-Rubaie, M.; and Chang, J. M. 2019. Privacy-preserving           Making 18(1): 1–10.
machine learning: Threats and solutions. IEEE Security &           Bize, R.; Volkmar, E.; Berrut, S.; Medico, D.; Balthasar, H.;
Privacy 17(2): 49–58.                                              Bodenmann, P.; and Makadon, H. J. 2011. Access to qual-
Amnesty International. n.d. Troll Patrol. https://decoders.        ity primary care for LGBT people. Revue Medicale Suisse
amnesty.org/projects/troll-patrol. Accessed: 2020-09-10.           7(307): 1712–1717.
Amnesty International Canada. 2015. LGBTI Rights. https:           Bjarnadottir, R. I.; Bockting, W.; Yoon, S.; and Dowding,
//www.amnesty.ca/our-work/issues/lgbti-rights. Accessed:           D. W. 2019. Nurse documentation of sexual orientation and
2020-09-10.                                                        gender identity in home healthcare: A text mining study.
                                                                   CIN: Computers, Informatics, Nursing 37(4): 213–221.
Andrus, M.; Spitzer, E.; Brown, J.; and Xiang, A. 2020.
‘What we can’t measure, we can’t understand’: Challenges           Bogen, M.; Rieke, A.; and Ahmed, S. 2020. Awareness in
to demographic data procurement in the pursuit of fairness.        practice: Tensions in access to sensitive attribute data for an-
arXiv preprint arXiv:2011.02282 1–12.                              tidiscrimination. In Proceedings of the 2020 Conference on
                                                                   Fairness, Accountability, and Transparency, 492–500.
Angwin, J.; Larson, J.; Mattu, S.; and Kirchner, L. 2016.
Machine bias: There’s software used across the coun-               Bolukbasi, T.; Chang, K.-W.; Zou, J. Y.; Saligrama, V.; and
try to predict future criminals. And it’s biased against           Kalai, A. T. 2016. Man is to computer programmer as
blacks. https://www.propublica.org/article/machine-bias-           woman is to homemaker? Debiasing word embeddings. In
risk-assessments-in-criminal-sentencing.                           Advances in Neural Information Processing Systems, 4349–
                                                                   4357.
Ashley, F. 2019. Gatekeeping hormone replacement therapy
                                                                   Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.;
for transgender patients is dehumanising. Journal of Medi-
                                                                   McMahan, H. B.; Patel, S.; Ramage, D.; Segal, A.; and
cal Ethics 45(7): 480–482.
                                                                   Seth, K. 2017. Practical secure aggregation for privacy-
Awan, I. 2014. Islamophobia and Twitter: A typology of on-         preserving machine learning. In Proceedings of the 2017
line hate against Muslims on social media. Policy & Internet       ACM SIGSAC Conference on Computer and Communica-
6(2): 133–150.                                                     tions Security, 1175–1191.
Balsam, K. F.; Molina, Y.; Beadnell, B.; Simoni, J.; and Wal-      Bosia, M. J.; McEvoy, S. M.; and Rahman, M. 2020. The
ters, K. 2011. Measuring multiple minority stress: the LGBT        Oxford Handbook of Global LGBT and Sexual Diversity Pol-
People of Color Microaggressions Scale. Cultural Diversity         itics. Oxford University Press.
and Ethnic Minority Psychology 17(2): 163.                         Bowleg, L. 2020. We’re not all in this together: On COVID-
Bamman, D.; O’Connor, B.; and Smith, N. 2012. Censor-              19, intersectionality, and structural inequality. American
ship and deletion practices in Chinese social media. First         Journal of Public Health 110(7): 917.
Monday .                                                           Brook. n.d.        Gender: A few definitions.           https://
Barocas, S.; Hardt, M.; and Narayanan, A. 2019. Fairness           www.brook.org.uk/your-life/gender-a-few-definitions/. Ac-
and Machine Learning. http://www.fairmlbook.org.                   cessed: 2020-10-01.
You can also read