小程序
传感搜
传感圈

What Humans Lose When AI Writes for Us

2023-10-06 14:54:01
关注

Artificial intelligence has pervaded much of our daily life, whether it’s in the form of scarily believable deepfakes, online news containing “written by AI” taglines or novel tools that could diagnose health conditions. It can feel like everything we do is run through some sort of software, interpreted by some mysterious program and kept on a server who knows where. When will the robots take over already? Have they already taken over?

The recent developments in AI offer existential questions we’ve been wrestling with since we put pen to proverbial paper: Who wrote this, and can I trust it? Fake news is old news, but some still argue over whether Shakespeare existed or represented multiple authors. Large language models (LLMs) are combinations of authors, each with their own style, voice and expertise. If the generative AI program ChatGPT keeps trying—and we keep feeding it Shakespeare—will it write our next great tragedy?

Linguist Naomi S. Baron of American University has been wading in the AI waters for years. In her latest book, Who Wrote This? How AI and the Lure of Efficiency Threaten Human Writing, she dives into the crux of the matter: If we hand over the written word to AI, what will we lose? Scientific American spoke with Baron on the issue of the ownership and trustworthiness of written communication now that AI is on the scene.

[An edited transcript of the interview follows.]

Did you use ChatGPT to write any of this book?

Sort of but just a smidge. I completed Who Wrote This? in mid-November 2022, two weeks before ChatGPT burst on the scene. It was a no-brainer that I needed to incorporate something about the new wonder bot.

My solution was to query ChatGPT about the intersection of this cutting-edge form of AI with issues such as creativity, education and copyright. In the book, I quote some of ChatGPT’s responses.

Credit: Stanford University Press

When I asked ChatGPT if it could hold copyright on short stories that it authored, the answer was “no” the first time I asked and “yes” the second. The discrepancy reflected the particular part of the dataset that the program dipped into. For the “no” answer, ChatGPT informed me that as an LLM, it was “not capable of holding copyrights or owning any form of intellectual property.”

By U.S. copyright law, that’s true. But for the “yes” response, the bot invoked other aspects of U.S. copyright: “In order for a work to be protected by copyright, it must be original and fixed in a tangible form, such as being written down or recorded. If a short story written by GPT meets these criteria, [ChatGPT said], then it would be eligible for copyright protection.

Consistency is the hobgoblin of large language models.

When thinking about AI-written news, is it all just a snake eating its own tail? Is AI writing just fodder to train other AIs on?

You’re right. The only thing relevant to a large language dataset is having text to consume. AI isn’t sentient, and it’s incapable of caring about the source.

But what happens to human communication when it’s my bot talking to your bot? Microsoft, Google and others are building out AI-infused e-mail functions that increasingly “read” what’s in our inbox and then draft replies for us. Today’s AI tools can learn your writing style and produce a reasonable facsimile of what you might have written yourself.

My concern is that it’s all too tempting to yield to such wiles in the name of saving time and minimizing effort. Whatever else makes us human, the ability to use words and grammar for expressing our thoughts and feelings is a critical chunk of that essence.

In your book, you write, “We domesticate technology.” But what does that “domestication” look like for AI?

Think about our canine companions. They descended from wolves, and it took many years, plus evolution, for some of their species to evolve into dogs, to be domesticated.

Social scientists talk about “domestication” of technology. Forty years ago personal computers were novelties. Now they’re ubiquitous, as are software programs running on them. Even Wikipedia—once seen as a dubious information source—has become domesticated.

We take editing tools such as spell-check and autocomplete and predictive texting for granted. The same goes for translation programs. What remains to be seen is how domesticated we will make text-generation programs, such as ChatGPT, that create documents out of whole virtual cloth.

How has your understanding of AI and LLMs changed how you read and approach writing?

What a difference three years makes! For my own writing, I remain old-fashioned. I sometimes still draft by hand. By contrast, in my role as a university professor, I’ve changed how I approach students’ written work. In years past I assumed the text was their own—not so today. With AI-infused editing and style programs such as Microsoft Editor or Grammarly, not to mention full-blown text-generation tools, at students’ beck and call, I no longer know who wrote what.

What are the AI programs that you feel are the least threatening, or that you think should be embraced?

AI’s writing ability is an incredible tour de force. But like the discovery of fire, we must figure out how best to harness it. Given the novelty of current programs, it will take at least several years to feel our way.

Today’s translation programs, while not perfect, are remarkably good, and the benefit is that everyday users who don’t know a language can get immediate access to documents they would have no other way of reading. Of course, a potential drawback is losing motivation for learning foreign languages.

Another promising use of generative AI is for editing human-generated text. I’m enthusiastic when AI becomes a pedagogical tool but less so when it simply mops up after the writer, with no lessons learned. It’s on users to be active participants in the composition process.

As you say in your book, there is a risk of valuing the speed and potential efficiency of ChatGPT over the development of human skills. With the benefit of spell-check, we can lose our own spelling proficiency. What do you think we’ll similarly lose first from ChatGPT’s ability to write legal documents, e-mails or even news articles?

As I argue in my book, the journalism business will likely feel the effects on employment numbers, though I’m not so much worried about the writing skills of the journalists who remain.

E-mails are a more nuanced story. On the one hand, if you use Microsoft Outlook or Gmail, you’ve already been seeing a lot of autocomplete when you write e-mails. On the other hand, the new versions of AI (think of GPT-4) are writing entire e-mails on their own. It can now literally be my bot writing to your bot. I worry that the likes of ChatGPT will lull us into not caring about crafting our own messages, in our own voice, with our own sentiments, when writing to people who are personally important to us.

What do you think of the recent and potential copyright infringement cases involving authors or publishers and ChatGPT?

The copyright infringement cases are interesting because we really are in uncharted territory. You’ll remember the case of the The Authors Guild v. Google, where the guild claimed Google Books enabled copyright infringement when it digitized books without permission and then displayed snippets. After many years of litigation, Google won ... under the ruling of fair use.

From what I’ve been reading from lawyers who are copyright experts, I suspect that OpenAI [the company that developed ChatGPT] will end up winning as well. But here’s the difference from the Authors Guild case: With Google Books, authors stood to lose royalties because users of Google Books were presumably less likely to purchase copies of the books themselves. With ChatGPT, however, if a user invokes the bot to generate a text, and then said user looks to sell that text for a profit, it could be a different ball game. This is the basis of cases in the world of generative art. It’s a brave new legal world.

参考译文
当人工智能为我们写作时,人类失去什么
人工智能已经渗透到我们日常生活的方方面面,无论是以令人毛骨悚然但逼真的深度伪造视频、带有“由AI撰写”标签的在线新闻,还是能够诊断健康状况的新型工具等形式出现。我们的一切行为都仿佛要通过某种软件进行处理,被某个神秘的程序解读,并存储在谁也不知道的服务器中。机器人究竟什么时候会接管一切?它们已经接管了吗?最近的人工智能进展带出了我们自执笔写下文字以来一直在思考的存在主义问题:是谁写了这些文字,我能信任它们吗?虚假新闻并非新鲜事,有些人至今还在争论莎士比亚是否真的存在,还是他代表了多位作者的合著成果。大型语言模型(LLMs)本身就是多个作者的集合,每个人都有自己独特的风格、语调和专业知识。如果生成型人工智能程序ChatGPT继续尝试——而我们继续喂它莎士比亚的作品——它会不会写出我们下一个伟大的悲剧?美国大学的语言学家诺米·S·巴隆(Naomi S. Baron)多年来一直在关注人工智能的发展。在她最新出版的著作《谁写了这部作品?:人工智能与效率诱惑如何威胁人类写作》中,她深入探讨了核心问题:如果我们把书写工作交给人工智能,我们会失去什么?科学美国人杂志(Scientific American)就人工智能时代书面沟通的所有权和可信度问题,采访了巴隆。[以下为采访的编辑版节录。] 你是不是用ChatGPT写了这本书的某些部分?某种程度上是的,但只是极少量。我在2022年11月中旬写完《谁写了这部作品?》,比ChatGPT正式亮相早了两周。很明显,我需要加入一些关于这个新奇AI工具的内容。我的解决办法是向ChatGPT提问,探讨这种前沿人工智能与创造力、教育和版权等问题之间的交叉点。在书中,我引用了ChatGPT的部分回答。图片来源:斯坦福大学出版社 当我问ChatGPT它是否可以拥有它自己撰写的小说版权时,第一次它回答“不能”,第二次却说“可以”。这种不一致反映了程序访问数据集的不同部分。在“不能”回答中,ChatGPT告诉我,作为一个大型语言模型,它“无法持有版权,也无法拥有任何形式的知识产权”,根据美国版权法,这是正确的。但在“可以”回答中,它又引用了美国版权法的其他方面:“为了获得版权保护,作品必须是原创的,并以有形方式固定下来,如书面或录音。如果由GPT撰写的小说满足这些条件,[ChatGPT表示],那么它就符合版权保护的条件。”大型语言模型的不一致性是它们的顽疾。谈到人工智能撰写的新闻,我们是否正在经历“蛇吃自己尾巴”的情况?人工智能撰写的内容是否仅仅是为了训练其他人工智能?你说得对。对于一个大型语言数据集来说,唯一重要的就是拥有文本可以“消化”。人工智能并不具备意识,对信息来源毫不在意。但如果是我们自己的机器人在和对方的机器人对话,那么人类交流又会发生什么变化?微软、谷歌等公司正在开发嵌入人工智能功能的电子邮件功能,这些功能越来越能“读取”我们的收件箱,并帮我们起草回复。如今的人工智能工具可以学习你的写作风格,并生成一个合理的仿制品,模拟你亲自撰写的内容。我的担忧在于,为了节省时间和减少努力,我们太容易屈服于这种便利。无论我们身上还保留着什么“人性”,能够使用词语和语法表达思想和情感的能力,都是我们本质中不可或缺的一部分。在你的书中,你写道:“我们驯化技术。”但就人工智能而言,这种“驯化”又是什么样子呢?想象一下我们的狗。它们的祖先来自狼,经过多年演变,才逐渐变成了我们今天熟悉的狗,被我们驯化。社会科学家谈到技术的“驯化”过程。四十年前,个人电脑还是新奇玩意,如今却已无处不在,运行在电脑上的软件也是如此。就连曾经被认为是不可靠信息来源的维基百科,也早已被我们所接纳。我们现在对拼写检查、自动补全和预测性打字等编辑工具都习以为常。翻译软件也是如此。有待观察的是,我们将如何“驯化”像ChatGPT这样的文本生成工具,它们可以凭空生成完整的文档。你对人工智能和大型语言模型的理解如何改变了你的阅读和写作方式?三年时间可真是天壤之别!就我自己的写作而言,我依然保持旧式习惯,有时还会手写草稿。相反,在我担任大学教授的职责中,我对学生写作的态度已发生了变化。过去我总是假定学生提交的文字是他们自己写的——但现在则不然。随着人工智能编辑和风格工具(如微软编辑器或Grammarly)以及更强大的文本生成工具的普及,学生随时都可以使用这些工具,我再也无法分辨哪些内容是他们自己写的。你认为哪些人工智能程序对你而言是威胁最小的,或者你认为我们应该接受哪些人工智能?人工智能的写作能力确实是一项令人惊叹的成就。但就像火的发现一样,我们也必须学会如何最好地利用它。鉴于当前人工智能工具的崭新性,我们至少需要几年时间来摸索适应。如今的翻译软件虽然不完美,但性能已经相当出色,而其价值在于,即使用户不懂某种语言,也能立即获取原本无法阅读的文档。当然,潜在的缺点是人们可能会失去学习外语的动力。另一个有希望的人工智能应用是用于编辑人类自己生成的文本。当人工智能成为教学工具时,我感到非常兴奋;但当它只是简单地帮人“收拾残局”、毫无学习过程时,我就没那么兴奋了。用户必须积极参与写作过程。正如你在书中所说,我们有风险把像ChatGPT这样的工具的效率和速度看得比培养人类技能更重要。有了拼写检查的帮助,我们可能会失去自己的拼写能力。你认为我们首先会因ChatGPT撰写法律文件、电子邮件,甚至新闻文章的能力而失去什么呢?正如我在书中所论述的那样,新闻行业很可能会在就业人数方面受到影响,但我对剩余记者的写作能力并不太担心。电子邮件则是一个更加复杂的议题。一方面,如果你使用微软Outlook或Gmail,你已经体验过写邮件时的自动补全功能。另一方面,像GPT-4这样的新一代人工智能已经可以自己撰写整封电子邮件了。现在可以真的由你的机器人写信给我的机器人。我担心的是,像ChatGPT这样的工具可能会让我们在写给对我们重要的人时,不再在乎用自己的语言表达自己的情感和想法。你如何看待近期以及未来可能涉及作者或出版商与ChatGPT的版权侵权案件?版权侵权案件很有趣,因为我们正处于前所未有的法律领域。你可以回忆一下“作家协会诉谷歌”一案,在该案中,作家协会指控谷歌图书未经授权数字化书籍并展示片段,构成版权侵权。在多年诉讼之后,谷歌胜诉……法院裁定这属于“合理使用”。从我读到的版权专家律师的观点来看,我推测OpenAI(开发ChatGPT的公司)最终也会胜诉。但与作家协会案件不同的是:在谷歌图书案中,作家们担心的是版税收入,因为谷歌图书的用户可能会因方便而减少购买纸质书籍的可能性。而在ChatGPT的情况下,如果用户使用AI生成文本,并试图将其出售牟利,情况可能就完全不同了。这正是生成艺术领域相关案件的基础。这是一个全新的法律世界,令人敬畏又充满挑战。
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

广告

scientific

这家伙很懒,什么描述也没留下

关注

点击进入下一篇

工业革命,从哪开革?

提取码
复制提取码
点击跳转至百度网盘