Abstract for Roundtable

Sara Mills defines a post-feminist text analysis as “one which is able to see that there are, within the parameters of the textual and discursive constraints, multiple interpretations of terms and discourses as a whole.” (Mills, 1998) Mills’ approach takes into account the impact of feminism and offers a more nuanced perspective compared to earlier analyses that focused on overt sexism and discrimination. Mills references Toolan’s argument that analysis should move away from “easy examples of sexism and racism” and to “subtler and hence more insidious discriminatory and exclusionary discourses that abound.”

Mills’ approach can be useful in evaluating, generating, and re-envisioning the machine-aided methods of text analysis that are becoming increasingly prevalent. In Lauren Klein’s book Data Feminism (2020), Klein offers a similar critique by comparing approaches to data by Facebook and the Make the Breast Pump Not Suck Hackathon. While Facebook expanded gender categories, their policies did little to protect and instead actively imperiled “the safety of some of the platform’s most marginalized users.” On the other hand, used reductive data categories to achieve goals of inclusion and equality. Klein’s examples demonstrate that a more nuanced approach to data and text analysis is necessary to to identify and address the subtler forms of discrimination and exclusion that can still occur.

In the field of translation, machine-assisted approaches are becoming more common. Although machine-assistance can improve efficiency and fidelity, it introduces another layer of complexity to the process of translation. Mills’ approach could aid in examining not only the product of translation but also the underlying systems and processes, including machine-assisted translation. While Taivalksoki-Shilov (2019) addresses some of the ethical concerns regarding machine-assisted translation such as the quality of translation and preservation of the author’s voice, other considerations are absent. Mills’ approach to textual analysis provides a useful framework for considering a broader range of perspectives and contextualizing editorial and linguistic choices within the cultural norms and ideologies of the source text.

Klein, Laura. What Gets Counted Counts. Data Feminism. MIT Press, 2020.

Mills, Sara. Post-feminist Text Analysis. Language and Literature, 1998, p. 241.

Taivalkoski-Shilov, Kristiina. Ethical issues regarding machine(-assisted) translation of literary texts. Perspectives 27.5 (2019): 689-703. DOI: 10.1080/0907676X.2018.1520907.

Rethinking textual data practices as feminist intervention

If, as Sara Ahmed writes, feminism is “a building project” (Living a Feminist Life, 14), then a critical point of feminist intervention in the textual analysis pipeline is in the selection, handling, and use of textual data. The text tokens — and the methods we use to parse, tag, and untangle them — are the nails and hammers, respectively, in that building project.

Every stage of identifying and preparing data (i.e., prior to its analysis) offers opportunities for feminist intervention. When working with an existing dataset (a readymade), we as researchers can interrogate its provenance, genealogy, potential for misuse, and more. Asking these questions can clue us into the assumptions that shaped its collection. 

If collecting data ourselves, we should look to domain experts to help us circumvent our own biases and situate knowledge within its local context — otherwise, we risk harming the communities the data claims to represent. When cleaning data, we have an opportunity to preserve difference and nuance by avoiding the tendency to aggressively standardize data to make it “easier” to analyze. 

We also have opportunities to rewrite or redistribute data in ways that actively oppose oppression — gender-based or otherwise — by revising or adding metadata, documenting and explaining datasets to promote more informed use, and even constructing alternative datasets that subvert dominant ideologies. These actions can not only bring our research efforts in line with a feminist aim, but also create the conditions for other researchers to do the same.

Book post: “#Hashtag Activism: Networks of Race and Gender Justice” by Sara Jackson, Moya Bailey and Brooke Foucault Welles.

Historical reference

Twitter was created in 2006 and since then it has transformed the way in which we communicate and engage with others. In 2007 Chris Messina suggested using the symbol “#” hashtag for the first time on Twitter, in order to organize and categorize related posts. The hashtag became an integral part of social media and then went on to spark some of the biggest political and social movements in recent years. In 2009 the hashtag #Iranelection became the first international-trending hashtag. The term “hashtag activism” appeared in 2011 in news reports to describe online activism with the hashtag branding. As of February 9, 2023, Twitter announced that it won’t be allowing access to its API. Currently only people with active accounts can still use the API…

Post

How can feminist activism use social media as a tool to challenge and reshape power structures and advocate for justice? In the book Hashtag activism the authors investigate some of the most relevant twitter hashtags addressing women’s violence, culture dynamics, discrimination of the trans community and racism. The authors focus on marginal communities that brought social issues situated on the periphery to a more central realm with the use of the hashtag, achieving a “considerable impact on political debates, social policy, and the public consciousness”. The creation of counterpublic spaces for the marginalized was seen then, as a win for inclusivity and visibility, in which whole networks benefited. 

The author’s intend not only to analyze popular hashtags individually like #metoo or #sayhername, but understanding the network activity between those tweeting about race, sexuality, and gender. They argue that “transforming online counterpublics into network data makes it possible to understand their dynamic and interdependent properties while also narrowing an overwhelming volume of data to a manageable size”. Allowing them to find patterns in how tweets achieve popularity and identifying which individuals within the networks become the main actors. Interestingly, each chapter contains an essay written by one influential person linked to the hashtags the authors selected.

The way in which the authors construct the networks is through the analysis of retweets and mentions (@). Once the network reaches a legible shape, the authors identify individuals, messages and subgroups of interest. They argue that their research benefits from using a mathematical agnostic approach, which relies on quantitative measures and statistical algorithms to identify relationships within a text and doesn’t require the researcher to have a preconceived notion, but just lets the data speak for itself. In addition, the authors investigated the mathematical properties of degree centrality and betweenness centrality, for network creation. Degree centrality measures the direct relationships between two nodes while the betweenness centrality is the measure of how often a node acts as a bridge. 

In order to reconcile digital tools and humanities, the authors employ Stephen Ramsey’s construct of algorithm criticism, which stresses the need to treat algorithms as cultural artifacts, set in social and political environments, to critically examine pre-assumptions and biases. Other methods, as feminist and critical race theories were also applied to investigate power and identity within the corpus. Finally, discourse analysis permitted the researchers to find meaning and significance beyond the popularity of certain hashtags and the social context in which these flourished to form attentional sets of networks. 

The first three chapters of the book explore women’s violence, black feminism and trans feminist advocacy, represented by trending Twitter hashtags like #yesallwomen #metoo, #sayhername, #youoksis and #girlslikeus. These hashtags have been effective at piercing the social media sphere by creating awareness and mobilizing the general public. The use of hashtags by social justice movements not only provided a means of communication but an opportunity to demand change, and influence the discourse in media news and political agendas.  

As computational methods are often used to study social media and identify patterns and trends, it is of extreme importance to understand the limitations and potential biases of each model as Richard Jean explains in his article “All Models are Wrong”. In contrast with this idea the authors used a series of methods to reduce bias, like agnostic modeling. However, agnostic models could in theory carry out mistakes as well if the data collected was incomplete or had already some form of bias. The authors recognize that text analysis is never perfect due to issues such as limitations at identifying relevant tweets, bias at analyzing only english-speaking populations instead of the global manifest, lack of context caused by the set number of characters and brevity of Twitter posts, or the potential for misinterpretation of sarcasm and irony in language. 

I consider that the authors were successful at taking a critical stand in their research and recognizing the limitations of it. One of their biggest contributions lies in the act of inviting the main voices representing the hashtags and correlating their experiences with their findings. This approach generates an extra layer of dimensionality that speaks to the social relevance of their work. Centering marginalized voices in the development and use of digital tools for social change as D’Ignazio stresses in her book Data Feminism. Another successful approach was not only to perform computerized data analysis but matching these to qualitative analysis, that contextualize and inform the first, offering a comprehensive view of the intersection between technology and social justice.

Yes, #hashtagactivism can suffer from the co-optation of corporate interests, the potential for online harassment and abuse, and even the limitations of achieving a wholesome sustained activism. But by exploring examples of successful instances in which the digital message permeates to the public to cause action and restructure frameworks of power, we can gain a deeper understanding of hashtag activism on social change. For now we’ll just have to work together to mitigate and advocate against the negative impact that modern technologies may bring. 

Building human context into how we understand and process textual data (response blog post to week 12 readings)

In the article “Automatically Processing Tweets from Gang-Involved Youth: Towards Detecting Loss and Aggression,” a team of computer scientists and social workers describe their use of natural language processing (NLP) to analyze tweets from known gang members in Chicago. Their goal is to understand when a tweet conveys a feeling of loss or aggression — which might precede an act of violence — with the hope that this automatic detection can support community outreach groups’ intervention efforts. 

The team realizes they cannot easily use existing NLP techniques given how significantly the textual data (i.e., the vernacular and slang found in the tweets) differ from the language these systems are trained on. Existing part-of-speech taggers, for instance, cannot parse the abbreviations and non-standard spellings found in the tweets. 

To create a more accurate model, the researchers hand-label the tweets, specifically turning to domain experts including social workers and two “informants” — 18-year-old African American men from a Chicago neighborhood with more frequent violence — to interpret each tweet. They call this act of decoding a “deep read” (2198), and the computer scientists work closely with the domain experts to understand the data and tune the model appropriately.

I feel this represents a feminist approach to working with textual data as it places primary importance on respecting the context of the data — and the people behind the data. Using existing NLP tools or foregoing the step of consulting domain experts would not only result in significant errors in the results, but could also misrepresent and even damage the communities the researchers are trying to support. As Lauren Klein and Catherine D’Ignazio write in Data Feminism, “Refusing to acknowledge context is a power play to avoid power. It’s a way to assert authoritativeness and mastery without being required to address the complexity of what the data actually represent” (ch. 6). 

Notably, the authors of the research article conclude by pointing to future research directions, explicitly stating that they aim to “extend our corpus to include more authors, more time periods, and greater geographical variation.” This suggests an iterative and process-oriented approach to modeling the data in line with Richard Jean So’s article “All Models Are Wrong,” in which he encourages researchers to assume that a model is wrong from the start but that it can be useful for exploration and can be improved through iteration (669). The researchers behind the Twitter data could (and should) continue to refine their model to better represent and engage with the people whose data they are modeling.

HDTMT: Text Mining Oral Histories in Historical Archaeology

Paper Title: Text Mining Oral Histories in Historical Archaeology

Author: Madeline Brown & Paul Shackel

Linkhttps://link.springer.com/article/10.1007/s10761-022-00680-5

Introduction:

Oral histories help provide important context for the past. It provides firsthand accounts from people who have experienced particular historical events in the course of their own life – a valuable insight for researchers to explore the social context, culture, politics, economics, etc. of the period being studied.

Text mining oral records and interviews is a relatively new method of retrieving important contextual information about the communities being studied in historical archaeology.

While the qualitative interpretation of texts typically conducted in historical archaeology is important and crucial, text mining and NLP can provide valuable insights that enhance and complement these methods.

The authors outline the following benefits:

(1) Rapid and efficient analysis of large volumes of data

(2) Reproducible workflows

(3) Reducing the potential for observer bias

(4) Structured analysis of subgroup differences

The authors also mention that they do not propose that text mining and NLP replace the traditional techniques used but as a supplementary tool.

This project not only shows an implementation of such techniques for historical archaeology, but it also provides a framework that can be reused and provides insights into the areas of improvement that can help further enable the use of text analysis and NLP methodologies for this particular area of study.

Data Source:

26 oral interviews acquired in 1973 and transcribed in 1981 from the anthracite coal mining region of Pennsylvania which were collected by the Pennsylvania Historical and Museum Commission and are now on file at the Pennsylvania State archives.

Link to Data Source:

Available on Github: https://github.com/maddiebrown/OralHistories

How Did They Make That:

– Interview transcripts that were available in PDF were converted to text format for analysis.

– Text was formatted to ensure that everything was standardized i.e., has the same sort of font, text size, etc.

– When PDF is converted to text sometimes extra characters and spaces come up which were removed to make the text tidy.

– Line breaks were added to signify speaker change.

– The text was then imported into R for further cleaning and then analysis.

– In R, metadata was removed. The text was then standardized. All terms were converted into lowercase (which for example avoids counting “coal” and “Coal” as separate terms), and numbers and apostrophes were removed.

– Tokens were then into individual words and bigrams, and stop words were removed “using tidytext’s stop_words lexicons: onix, SMART, and snowball.”

– After this certain slang terms, names, and common words which did not influence any meaning of the text were removed.

– In terms of actual text analysis, in this project, the authors focused on word frequency and n-grams, and they discussed future potential of sentiment analysis and tagged lexicon analysis.

Tools/skills you would need to do the project yourself:

– R Programming Language

– Text mining and NLP concepts (for example what stop words to remove and why)

– R Packages Used: tidyverse, textclean, tidytext, and cleanNLP

Tutorials:
(I used Google search to come up with primary resources, although interested analysts are suggested to search more resources for further understanding)

– R: https https://www.datacamp.com/courses/free-introduction-to-r

– Text mining with R: https://www.tidytextmining.com

– NLP with R: https://s-ai-f.github.io/Natural-Language-Processing/

– tidyverse R package: https://www.tidyverse.org/learn/

– textclean R package: https://cran.r-project.org/web/packages/textclean/index.html

– tidytext R package: https://cran.r-project.org/web/packages/tidytext/index.html

– cleanNLP R package: https://cran.r-project.org/web/packages/cleanNLP/index.html

Response Blog Post – Week 9

AI Ethics: Identifying Your Ethnicity and Gender

This reading showcases the use of NLP and Machine Learning to build an AI model that can accurately identify gender based on a single tweet.

But it begins with the questions around the ethics of allowing AI to detect ethnicity and gender, followed up with more questions and then some more and before the author even tackles the pros and cons properly dives straight into building a model to detect gender from a tweet.

The article then showcases how that is achieved, how the data is cleaned and processed, how the model is built and finally applied to showcase its accuracy. However even the model itself leaves more questions than answers like what percentage of model accuracy is acceptable? Are tweets from celebrities the best training dataset to predict gender for everyone or just a particular niche of people?

But coming back to the ethical question of allowing AI to predict gender or ethnicity, it can be very unethical and harmful in many ways in my personal opinion and here is one example I had in mind.

Let’s say you are applying for a job application with your resume and cover letter and the employers run it through an AI model. Now the AI model they have has been trained on their own dataset of employees and tries to detect those who are likely to be more “successful” at their jobs. Consider this to be in the finance industry which is dominated by mostly white men. Since, their AI model has been trained on a dataset that in itself has coded bias because it is based on a training dataset that itself is biased on both ethnicity and gender, this can cause a problem if somebody with a different set of ethnicity and gender applies (for example a non-binary Hispanic person). This is why I personally would not advocate using AI to detect ethnicity and gender unless the training dataset itself is adequately representative of a rather diverse demographic or there is an absolute need.

Beyond Fact-Checking: Lexical patterns as Lie Detectors in Donald Trump’s Tweets

As the title suggests, this paper is on figuring out a model to detect lies in Donald Trump’s tweets.

But why does that matter and why can’t we simply label things as lies in a journalism context? Isn’t fact just different from opinions / lies / misinformation? Because, as the author explains, “the line between factual claims and statements of opinion can be difficult to draw” (Graves, 2016, p. 92). There can be “opinionated factual statements” where the statement is indeed factual but draws on a substance of opinion, misinformation as questions where while you are directly not making a misinformed statement but ask a question that would lead to a thought of misinformation, and weaving misinformation into “otherwise truthful statements” (Clementson, 2016, p. 253) where the statement you are making it true but needs more context as without it misinformation is propagated.

What becomes an even greater difficulty is understanding not only whether a statement is fact or false but if it is false, was it an “honest mistake or strategic misrepresentation”. There are debates on this questions itself whether and exactly when in journalism you should present something as an intentional lie.

The research paper’s primary aim is to see if any language patterns emerge in Donald Trump’s tweets to identify false information. It aims to do this by comparing the language of his tweets when he shares true and false information and find the lexical patterns in each to create a model for detection.

They focused on some patterns that are associated with lying in language like the use of words to express a negative emotion, certain pronouns etc. The study then found that indeed a lot of Trump’s tweets that had false or misleading information actually did have a high number of such patterns existing. This can be used in journalism and scholarship beyond simple fact checking when needed.

Gendered Language in Teacher Reviews

This is an interactive visualization with a chart that allows exploration of words used to describe male and female teachers from a dataset of reviews from RateMyProfessor.com. A word or two-word phrase can be entered, and the chart generates to show the usage of words (per million words of text) broken down by gender (male & female) which can be further broken down into “positive” or “negative” ratings which enables further exploration to see in what context the word was being used.

So, basically when a word or two-word phrase is typed in, it crawls the database of reviews from RateMyProfessor.com, searches up how many times those words have been used per million words of text and breaks it down by gender and further breaking it down by positive or negative review and creates a chart from it while showing which department / subject the professor teaches.

I personally think the visualization does match the intention and carries out the task quite swiftly. The data visualized matches the question asked in at least its literal meaning.

For me I also particularly liked the break down by department as well. For example, searching for “programming” shows computer science followed by engineering as the ones for which the word shows up the highest while “aesthetic” shows up the most for fine arts.

Also, it made me stop and think once that while we are trying to find out how certain words can have a gender bias in reviews, I as the prompter typing in words am making decisions on what words I am choosing that I think will have a gender bias (making me think of my own biases in gender language).

Film Dialogue

This project comes in a few parts to it.

The first analyses 30 Disney films and breaks down the screenplay by gender which showed that 22 out of 30 of those films have males dominating dialogues even in a film like Mulan where the leading character is female.

The second visualization contains a larger dataset of 2000 screenplays where the researchers’ linked characters with at least 100 words of dialogue to their respective IMDB pages which can showcase gender of the actor. This is then used to visualize screenplays on a scale of 100% of Words are Male to 100% of Words are Female and the chart is interactive so you can hover over one to see the name of the film and the breakdown of dialogue percentage by gender.

The third visualization takes into consideration only high-grossing films defined as films rankings in the top 2,500 of the US Box Office. It shows a general view of a scale from 100% male to 100% female like before. But it does a better job of showing just how less of the films have female dominated speech or even a 50-50 gender balance.

In this visualization, you can explore further by clicking on a film title and it shows the exact percentage breakdown of dialogue by gender, who the top 5 characters are (by number of words) and a breakdown of the male or female lines by minutes of the film from start to end showing the increase (or decrease) of each against the timeline of the film.

The fourth visualization shows a bar chart of the percent of dialogues by actor’s age broken down by gender. This shows that women actors have less dialogues available when they are over the age of 40 while for men it is the exact opposite, and they have more roles available with age.

The last visualization is a more interactive one where one can search up a particular film, see its gender dialogue breakdown and number of words by characters. There are also some filtering options like years and genres.

The visualizations do look really cool overall, and I just loved how clean and aesthetic everything is.

Gender Analyzer

This is a tool where one can enter certain text and it tries to analyze whether the text was written by a man or a woman. I frankly do not know how good this tool is and have been very confused as I have tried to put in text from woman writers which have been labeled as masculine and text from male writers labeled as feminine. It does say that it has an accuracy of about 70% which means it is not that perfect. But I just did not understand the use case or like why you even need this in the context it is supposed to be used.

Response Blog Post – Week 8

My key takeaway from this week’s readings is summarized by Catherine D’Ignazio and Lauren Klein’s chapter title in Data Feminism “What Gets Counted Counts”. Public policy and decision making at regional to national to international level takes place in the modern-day world based on conclusions drawn from the data being collected. And what is in that data matters because that dictates what takeaways we have and how we better understand the world around us.

Being counted is the equivalent of having representation in the data age. When census surveys are done for example, allocation of resources and even the amount of representation a region gets is based on the results of the surveys conducted. This poses a problem in cities like New York City where there is a significant undocumented immigrant population, and the census fails to properly “count” them which leads to a lack of resources when compared to the actual population that such regions have.

In fact, during the last census drive, I had briefly volunteered with a community organization that was helping collect the survey forms and I have seen in person how many people in the community, particular undocumented immigrants and those seeking asylum do not want to fill up the survey in fears that it makes them vulnerable. This creates a problem because most of these people are already from a demographic that is underrepresented in politics and without being represented on the census, they are only going to be more underrepresented.

I also really liked the point that sometimes lack of data or missing data is representative of a situation itself. This is best described by the Guardian interactive “Does the New Congress Reflect You” where users can select their demographic and see how many people like them are in Congress. Clicking on “trans + nonbinary” leads to a blank map showing that there are no people in Congress like them. That is powerful as it shows the lack of representation for trans & nonbinary folks in the 2018 Congress.

Overall, what the chapter presents is the fact that data is not impartial and rather has biases and inequalities embedded in the whole process from collection, analysis and the conclusions drawn which in turn perpetuates the said biases and inequalities. Hence as data feminists, it is important to be able to consider the misrepresentations and injustices that are embedded in the data process to confront and tweak these issues.

Book Report & Review Blog Post_YW

Weapons of Math Destruction – Cathy O’Neil

a.) summarizes the main takeaways of the book for classmates who have not had the opportunity to read it

Cathy O’Neil’s book “Weapons of Math Destruction” examines how mathematical models and algorithms are frequently employed to support and maintain systemic injustice and inequality. According to O’Neil, these “Weapons of Math Destruction” (WMDs) have serious adverse effects on people and society as a whole, especially on vulnerable and marginalized groups. The book offers numerous examples of WMDs being used in various fields, including hiring, advertising, education, and criminal justice. O’Neil, for instance, demonstrates how discriminatory and unfair predictive models may be when used to assess a teacher’s performance or a person’s creditworthiness.

The key lessons to be learned from the book include the necessity for increased accountability and openness in the design and use of algorithms, as well as the significance of adding moral considerations into algorithmic decision-making. The book also emphasizes the potential for algorithms to reinforce and magnify prejudices, highlighting the significance of diversity and inclusion in the technology sector.

In the Introduction to the book, the author sets the stage for her argument by describing how algorithms are increasingly being used to make decisions that have significant consequences for people’s lives, such as who gets hired or fired, who gets a loan or a mortgage, and who gets sent to prison. It notes that these algorithms are often proprietary and secret, meaning that the people affected by them have no way of knowing how they work or challenging their decisions.

There’s a popular saying that “men lie, women lie, but numbers don’t.” Because people tend to believe that ‘numbers don’t lie,’ many tend to cow into submission to anything that is based on numbers. However in “Weapons of Math Destruction” the author proves the use of warped mathematical and statistical models in algorithms to influence against ordinary people. These ‘algorithmic’ decisions tend to entrench existing inequalities by empowering the rich and powerful against the helpless mases. He debunks the notion of algorithmic neutrality with the argument that algorithms are based on data which are obtained from recorded behaviors and choices of people – most of which are flawed.

The author confirms Kate Crawford’s observation of the use of obscuring by mystification to conceal the truth from the people affected thereby. When confronted, computer scientists tend to present answers which suggest that the internal operations of the algorithms is ‘unknowable’ thereby slamming the door against all questioning. In line with the theory of political economy, the author observes that the effectiveness of algorithms is evaluated based on its ability to bring in the currency: i.e. political power to politicians, and money to business, but never on its effect on the affected people. Examples include the use of value-added modelling against teachers and scheduling software to optimize profits while exploiting desperate people and exacerbating working conditions of the worker and their social life. Political microtargeting which undermines democracy and provides politicians an avenue to be elusive by being ‘many things to many people.

b.) connects the book to our class conversations:

The book makes various connections to our class discussions on Feminist and Feminist Text Analysis. First, it draws attention to the ways in which algorithms can perpetuate systemic prejudices and discrimination, which can have serious adverse effects on disadvantaged and vulnerable communities, including women. Second, the book emphasizes the significance of including other viewpoints and voices in algorithmic decision-making processes, which is consistent with the intersectionality tenet of feminism. Finally, the book advocates for algorithmic decision-making to be more open and accountable, which is crucial for guaranteeing fairness and equity for all people, particularly women and other underrepresented groups.

c.) suggests what perspectives or new avenues of research and thought the book adds to the landscape of computational text analysis.

The book expands the field of computational text analysis by introducing a number of fresh viewpoints and lines of inquiry. One of the book’s major contributions is focusing light on the unfair application of mathematical models and algorithms in decision-making procedures like hiring, lending, and criminal justice, which can have a big impact on people’s lives. The book casts doubt on the idea of algorithmic neutrality by demonstrating how algorithms are built on inaccurate data derived from observed human actions and decisions, producing biased results that frequently worsen already-existing disparities.

Moreover, the impact of algorithmic decision-making on people, which reduces them to insignificant numbers and ignores their personal histories, psychological conditions, and interpersonal interactions, is highlighted in the book. It exposes the potential biases and inequities inherent in algorithmic judgments and emphasizes the necessity to address the ethical implications of using only algorithms to analyze human tales.

 Since many algorithms employed in significant decision-making processes are private and secret, it can be challenging for those who are affected by these judgments to understand how they operate or to contest them. This is why the book examines the topic of transparency and accountability in algorithmic decision-making. The book highlights the need for greater accountability and transparency in the creation and usage of algorithms and urges readers to evaluate these tools’ effects on society with greater knowledge and critical thought. The use of computational text analysis in domains like education, where algorithms are employed to evaluate professors and lecturers, and the potential biases and limitations of such evaluations are also raised in the book. It promotes deeper study and reflection on the creation of moral and just algorithms that take into account the intricate social and cultural influences on text data and analysis.

d.) Own critical reflections

In the chapter 7- Sweating Bullets, the author highlights an important issue: the unfair use of past records and WMD (weapons of math destruction) to screen job candidates, resulting in the blacklisting of some and the disregard of many others. When we rely on an algorithmic product to analyze human stories, individuals become mere numbers. For instance, a hard-working teacher’s efforts for a day are reduced to 8 hours in the database. Similarly, the practice of Clepening operates on the same principle. The machine does not care about an individual’s mental stress, personal preferences, or relationships; it only considers the additional hours worked. The Cataphora software system operates in the same manner. During the 2008 recession, companies used the software’s decision to lay off employees who had small and dim circles on the chart. While I agree with most of the author’s statements, I remain optimistic that with advancements in AI, the damages caused by WMD can be reduced. Although I am unsure of how this can be achieved, the author has addressed many of the problems, and solutions may exist. This chapter provided an example of Tim Clifford’s teacher evaluation case, it reminded me of the Student Evaluation of Teaching that is conducted every semester at City Tech, as well as at all other CUNY undergraduate colleges. These evaluations allow students to provide feedback on their classes before the final exams to eliminate potential bias. The feedback is then gathered and analyzed to help instructors improve their teaching. Prior to the pandemic, City Tech used a paper version of the evaluations, where professors would receive forms for each class and ask students to fill them out in class. Instructors had to leave the room while students filled out the forms, and a student would then send the completed forms to the Assessment and Research office. However, this evaluation process put pressure on some instructors, particularly those who were adjunct professors or had not yet received tenure. Some instructors chose to not distribute the forms to students, filled out the forms and submitted forms themselves etc. Despite the potential for bias from students, I believe that the Student Evaluation of Teaching questions are reasonable and can help instructors improve their teaching methods. At the same time, I recognize that the evaluation process may not be entirely fair to instructors, and that algorithms used to evaluate teaching may also be subject to biases and inequalities. Therefore, it is crucial to prioritize the development of ethical and fair algorithms that account for the biases and inequalities present in our society.

Respnse blog post_Week 4_2.27.23_YWei

Post-feminist text analysis by Sara Mill

Speaking in Tongues: Dialogics and Dialectics and The Black Woman Writer’s Literary Tradition

The author analyzes popular cultural texts using post-feminist theories. She investigates how gender and power are represented in these texts and how they reflect and influence societal beliefs and values about gender.

Sara Mills’ post-feminist text analysis work is relevant to the course goal of learning feminist text analysis because it highlights the ongoing struggle for gender equality in language and discourse. She also demonstrates how language is used to subtly undermine feminist goals and promote traditional gender roles by analyzing popular media texts such as advertisements. She demonstrates, for example, how women are frequently objectified and reduced to their physical appearance, whereas men are portrayed as powerful and dominant. And we have many scholars investigating how language constructs and reinforces gender roles, stereotypes, and power dynamics in feminist text analysis. Mills’ research expands on this foundation by investigating how postfeminist discourses that claim to have achieved gender equality actually perpetuate sexist attitudes and limit women’s agency.

By contrast, Mae G. Henderson’s article emphasizes how black women writers use language to challenge dominant cultural narratives and give voice to marginalized perspectives. Though Henderson’s article is not explicitly feminist, it can be viewed as part of a broader feminist project that seeks to amplify marginalized voices and challenge dominant cultural discourses.

In both articles, the importance of examining how language and media representations perpetuate stereotypes and power imbalances is stressed. In addition, they emphasize the importance of diverse representations that reflect marginalized perspectives and experiences. Both articles engage in feminist text analysis in critiques and challenges of dominant cultural narratives and promote greater diversity and inclusivity in media and literature.

Response blog post_Week 2 _2.6.23_YW

Sex and gender are often separated because they refer to distinct aspects of a person’s identity. gender is a social construct that can vary across cultures and can change over time. By separating the two concepts, it is possible to understand and address the ways in which gender and sex intersect and how they impact an individual’s experiences and opportunities. As we discussed in class today, gender is the question that is constantly asked when we need to fill out forms or join up for something. This reminds me of all the data analysis we did at City Tech for student enrollment, graduation, and retention, or surveys we hand out to gather data. We always included the variable of gender in our analyses. I thought it was interesting that the most recent Enrollment Dashboard, which we just updated for Spring 2022, had five demographic factors under the gender category: Men, Women, Non-binary Persons, Gender Nonconforming Persons, and Unspecified. We only have the gender categories of Men and Women when we get the data from the CUNY IRDB database prior to Spring 2022. This shift towards a more inclusive understanding of gender has led to an increase in the number of variables for gender in data analysis, allowing for a more nuanced and accurate representation of gender identities. And I believe it’s important for ensuring that data analysis is inclusive and respectful of all gender identities, and for providing a more complete picture of the experiences and perspectives of individuals who identify outside of the male/female binary.