Day 25 - In-Class - Data Dilemmas: Ethics in an Algorithmic World#

✅ Put your name here

#

✅ Put your group member names here.

#

Learning Goals:#

By the end of this assignment, you should be able to:

  • Evaluate the impact the data and algorithmic bias can have on real-life decisions

  • Differentiate between intent and impact of Machine Learning models

  • Identify benefits and consequences of using ML and AI tools

  • Compare similarities between models and simulations to real-life inequities

  • Read and interpret professional code built to create a simulation

Algorithms in our Lives:#

✅  As a group, create a list of algorithms and models that you interact with on a daily basis. Are they determinstic or stochastic? What bias is present in these algorithms? How do you know? Be prepared to share an example you came up with to the class.

Write down your answers here.

Survial of the Best Fit#

In the pre-class assignment, you played “Survival of the Best Fit” and explored how bias inherent in our world and data can influence the results of an algorithm. Today, we are going to take a deeper dive in to the data that was used to create the simulation and explore how this game mimics hiring practices used by everyday companies.

✅  In your groups, take a few minutes to discuss things you noticed about the game. Write down a few thoughts below.

Write down some notes here.

Taking a Deep Dive into the Code#

Now open the biased_data_gen.ipynb file from the website. As a group work through that notebook.

✅  After you finish going through the biased data generation notebook, take a few minutes to reflect on the code and the data. Write down some thoughts below. How does the simulated data reflect real systemic issues present in our world?

Write down your answers here.

Real Life Application#

It turns out Amazon (and many more companies more recently) have been using Machine Learning and AI algorithm to assist in their hiring practices with mixed success.

✅  Take a few minutes to read “Amazon ditched AI recruiting tool that favored men for technical jobs”.

Discuss#

✅  Discuss the follow questions with your group and take some notes on what you talk about.

  1. What connections do you seen between “Survival of the Best Fit” and Amazon’s real-life recruiting tool?

  2. Why was Amazon’s algorithm bias? What was biased about the training data?

  3. How would you feel if you knew a company you applied to was using AI to judge candidates’ applications?

  4. If you were in a hiring role, would you feel comfortable using AI in hiring decisions? What steps might you take to help mitigate bias?

Write down some notes here.

Different Platforms, Different Algorithms#

Now that you have explored how bias can be embedded in data and algorithmic choices, let’s take a look at how different algorithms can be used to shape user’s experiences on different platforms. Specifically, let’s take a look Twitter (now called X) and BlueSky. Your group will split in half, with one half researching Twitter and the other half researching BlueSky.

✅  Take notes below about what you learn about how each platform uses algorithms to shape user experiences. Make sure to include the references or links to the sources you used to find this information. Be prepared to share at least one finding with the class.

Write down some notes here.

Share Your Highlights#

We will spend the last 10-15 minutes of class sharing what you learned and discussed. Make sure to note a few things to share with the rest of the class.

Takeaways from Today’s In-Class#

Today’s assignment should have made you more aware about the power and the real-life harm that algorithm can cause. While they can provide great benefits like being cost-effective and speeding up time-consuming processes, they can also perpetuate harm existent in our world. Some more takeaways:

  • Pros:

    • Algorithms can help reduce workload and synthesize large amounts of data.

    • When implemented appropriately and thoughtfully, algorithms may help uncover patterns in candidates.

  • Cons:

    • Algorithms reflect the data they are given. Bias in the data implies there is bias in the model.

    • As we learned earlier in the class, all data contains bias!

    • Bias goes beyond gender, as we talked about in this case. It can also portray ageism, ableism, racism, socioeconomic status, etc. Similarly, algorithms can favor key words in resumes that might not be present in resumes of candidates that are highly qualified.

    • Not using data that explicitly contains race, sex, or other identities does not mean that our algorithms will not negatively impact certain groups. These groups tend to be the most marginalized and underrepresented.

Read More in Algorithmic and Data Bias.

Add any last reflections here.


Assignment wrapup#

Please fill out the form that appears when you run the code below. You must completely fill this out in order to receive credit for the assignment!

from IPython.display import HTML
HTML(
"""
<iframe 
	src="https://cmse.msu.edu/cmse201-ic-survey" 
	width="800px" 
	height="600px" 
	frameborder="0" 
	marginheight="0" 
	marginwidth="0">
	Loading...
</iframe>
"""
)

Congratulations, you’re done!#

Submit this assignment by uploading your notebook to the course web page.

See you next class!

This assignment was designed by Emily Bolger and Rachel Roca (2024).

© Copyright 2024, Department of Computational Mathematics, Science and Engineering at Michigan State University, All rights reserved.