Tag: citations

  • Generative AI for College Students Series: Baked In Biases in AI Tools Reflect and Amplify Prejudices

    Image created with Stable Diffusion.

    Please keep in mind that new technology like Generative AI (Gen AI) shouldn’t simply make your thinking or work easier, much less take the place of the uniquely singular abilities of human beings to grow cognitively, think creatively, or evaluate critically. If you use Gen AI to simply avoid work, you are doing it wrong. Instead, using Gen AI in the spirit of Douglas Engelbart’s “augmenting human intelligence” and Donna Haraway’s configuration of the cyborg point the way to beneficial heightening of human possibility instead of harmful erasure of the cognitive distinctions of humanity. If you use Gen AI, use it wisely and use it well. This post is the ninth in this series.

    Science and technology are not neutral and bias free. While we might aim to elevate them above human biases, they are part of human culture and carry the weight of the best and worst of ourselves. Similarly, generative AI tools, trained on vast datasets that reflect the biases of society, can reproduce and amplify these distortions in their responses. This raises important questions about the role of AI in academic writing and the potential for perpetuating prejudice.

    AI tools are not neutral; they reflect the biases present in their training data. For example, if a dataset contains stereotypical portrayals of certain groups, the AI will likely reproduce those stereotypes in its responses. This can result in biased or offensive content that undermines the credibility of a student’s work. Moreover, because AI-generated text is often polished and coherent, students may be less likely to question its content, thereby unintentionally perpetuating harmful ideas.

    Consider a student writing a paper on gender roles in society. They prompt an AI tool to provide an analysis, and the AI responds with a well-written paragraph that reinforces outdated stereotypes. The student, assuming the AI is neutral, incorporates this analysis into their paper, potentially spreading biased ideas. This scenario highlights the danger of relying on AI without critically evaluating its responses.

    It bears noting that while we influence the development of technology, it in turn influences human culture. In the case of AI tools, the technology is not only shaped by society but also actively reshapes it by amplifying existing biases. Students must recognize this dynamic and take steps to mitigate its impact on their work.

    To address this problem, students should actively seek out diverse perspectives and critically evaluate AI-generated content. They can do this by comparing AI responses to credible sources and looking for inconsistencies or biases. Additionally, educators can play a crucial role by teaching students to recognize and challenge biases in AI-generated text. This might involve incorporating discussions of AI bias into the curriculum and providing tools for analyzing and addressing it.

    While AI tools can be valuable writing assistants, they are not immune to the biases of the data they are trained on. The cyborg student must approach these tools with a critical eye, recognizing the potential for bias and taking steps to mitigate its impact. By doing so, they can produce work that is not only well-written but also equitable and inclusive.

  • Generative AI for College Students Series: Watch Out for Fabricated Footnotes and Fake Citations

    an anthropomorphic cat as a professor wearing a suit and orange tie standing in front of a chalkboard in a classroom
    Image created with Stable Diffusion.

    Please keep in mind that new technology like Generative AI (Gen AI) shouldn’t simply make your thinking or work easier, much less take the place of the uniquely singular abilities of human beings to grow cognitively, think creatively, or evaluate critically. If you use Gen AI to simply avoid work, you are doing it wrong. Instead, using Gen AI in the spirit of Douglas Engelbart’s “augmenting human intelligence” and Donna Haraway’s configuration of the cyborg point the way to beneficial heightening of human possibility instead of harmful erasure of the cognitive distinctions of humanity. If you use Gen AI, use it wisely and use it well. This post is the eighth in this series.

    In the science fiction film Blade Runner, replicants—advanced androids indistinguishable from humans—question the nature of their existence. Similarly, students using generative AI tools to write papers may find themselves grappling with questions of authenticity, particularly when it comes to citations. While AI can generate well-formatted citations and quotes, these may be entirely fabricated, leading to academic dishonesty and intellectual confusion.

    The problem arises because AI tools do not “know” the sources they cite. They generate citations based on patterns in their training data, which may include errors, inaccuracies, or outright fabrications. For example, an AI might invent a book title, attribute a quote to a nonexistent author, or misrepresent the content of a real source. These fabrications can be subtle and difficult to detect, even for experienced scholars.

    Imagine a student writing a paper on the ethics of artificial intelligence. They prompt an AI tool to include a quote from a prominent philosopher. The AI responds with a quote that seems relevant and includes a properly formatted citation. However, when the student checks the source, they discover that the philosopher never wrote those words, or that the book cited does not exist. This scenario is not only frustrating but also undermines the integrity of the student’s work.

    This issue mirrors the theme of simulacra in science fiction—copies without originals. In Jean Baudrillard’s theory of simulacra, representations of reality become more important than reality itself. AI-generated citations are simulacra of academic integrity, creating a false appearance of legitimacy. Just as replicants in Blade Runner question their humanity, students must question the authenticity of AI-generated citations.

    To combat this problem, students must adopt a cautious approach to AI-generated citations. They should avoid prompting AI tools to generate citations outright and instead use AI to assist with finding credible sources. For example, a student could ask the AI to suggest relevant authors or topics, then locate and verify those sources independently. This approach ensures that the citations are accurate and legitimate.

    In conclusion, while AI tools can be powerful assistants, they are not substitutes for human judgment and critical thinking. The cyborg student must learn to use these tools selectively, always prioritizing accuracy and authenticity. By doing so, they can maintain the integrity of their academic work and avoid the dangers of fabricated footnotes.