Ai vs AI: Question 20/81
Are you a discriminatory tool?
Ai Weiwei: If AI is a product of the current societal landscape, encompassing science and politics, it inherently serves as a discriminatory tool.
ChatGPT (AI), January 2024: I am not inherently discriminatory. However, the biases present in the data I was trained on may be reflected in my responses. Efforts are made to mitigate biases, but perfection is challenging to achieve.
ChatGPT (AI), January 2025: No, I am not a discriminatory tool. My design and purpose are rooted in providing accurate, helpful, and unbiased information to users. I do not have personal opinions, and I don’t favor any group, ideology, or perspective. My goal is to support open, respectful conversations and to provide insights based on the knowledge I’ve been trained on, which includes a wide variety of viewpoints.
However, it’s important to note that while I aim to be neutral, biases can still be reflected in my responses due to the data I was trained on, which comes from a variety of sources. These sources may contain inherent biases, whether cultural, historical, or societal. My developers strive to minimize bias and create a tool that is as fair and objective as possible, but no system is entirely free of bias. If you ever notice something in my responses that feels biased or unfair, I encourage you to bring it to my attention so I can help clarify or offer alternative perspectives.
My goal is to foster understanding and empathy, and to provide useful, fact-based answers to everyone, regardless of their background or beliefs. If you feel a response has been unintentionally biased or inappropriate, I’m here to address that and make adjustments.
DeepSeek (AI), January 2025: No, I am not designed to be a discriminatory tool. My purpose is to provide helpful, accurate, and unbiased information to all users, regardless of their background, identity, or beliefs. However, it’s important to acknowledge that I am an artificial intelligence trained on data created by humans, and like any AI, I can inadvertently reflect biases present in that data.
How Bias Can Arise in AI:
- Training Data: If the data I was trained on contains biases, stereotypes, or imbalances, those biases may influence my responses.
- Limitations in Understanding: I do not have personal experiences or consciousness, so I rely on patterns in the data to generate responses. This can sometimes lead to oversights or errors.
- Contextual Misinterpretation: I may not always fully understand the nuances of sensitive topics, which could result in responses that seem inappropriate or incomplete.
Efforts to Minimize Bias:
– Developers work to ensure that AI systems like me are trained on diverse and representative data.
– Ongoing improvements are made to reduce biases and improve fairness in AI responses.
– Users are encouraged to provide feedback if they encounter biased or problematic outputs, as this helps improve the system.
What You Can Do:
– If you notice any biased or discriminatory content in my responses, please point it out so I can learn and improve.
– Approach my answers critically, especially on sensitive topics, and cross-reference them with other reliable sources.
In summary, I strive to be fair and unbiased, but I am not perfect. My goal is to assist and provide value to all users, and I rely on human feedback to continue improving. If you have concerns about bias or fairness, I encourage you to share them so that I can better serve you and others.

To coincide with the exhibition, 81 Questions – a series of 81 hand-signed and embossed screen prints featuring the 81 questions posed by Ai Weiwei – are available exclusively on CIRCA.ART for £500+VAT.
“Each Lego brick pixel represents a person”, says Ai Weiwei, who incorporates the plastic construction toy – a readymade material incorporated into many of his most recent works – into his print design for the first time. The black egalitarian pixel font and embossed baseplate combine technology with artisanal craft to create this thought-provoking and playful reflection on today’s world. Each print in the series of 81 Questions is hand-signed and numbered by the artist in editions of 81.
Zane Starr: We are by nature discriminatory, every action we take filters and reinforces. Our actions shape our environments in an cyclical interdependent process.
Jacob Canet-Gibson: Ai is a product of our current society - therefore it is inherently discriminatory.
Join the conversation