小程序
传感搜
传感圈

Secret Messages Can Hide in AI-Generated Media

2023-06-24 16:12:08
关注

On June 27, 2010, the FBI arrested 10 Russian spies who lived and worked as American professionals near New York City. The case, which unraveled an intricate system of false identities and clandestine meetings, exposed one of the largest spy networks in the U.S. since the Cold War ended and inspired the show The Americans.

It also brought attention to steganography, a way of disguising a secret message within another message. The New York spies hid their secrets in plain sight, encoding communications within the pixels of seemingly innocuous images posted on publicly available websites. To read them, the recipient had to download an image, translate it into the 1s and 0s of binary code, and know which altered digits, taken in sequence, would spell out the secret. 

Steganography, which is both an art and a science, differs from the better-known method of secret communication known as cryptography. Where cryptography intentionally conceals the content of a message, transforming it into a tangle of text or numbers, steganography conceals the fact that a secret exists at all. “Steganography hides the presence of the message,” said Christian Cachin, a computer scientist and cryptographer at the University of Bern. “If an adversary can detect a hidden message, then the sender has lost the game.” 

As with any method of covert communication, the challenge is how to make it perfectly secure, meaning neither a human nor a machine detector would suspect a message of hiding a secret. For steganography, this has long been a theoretical possibility, but it was deemed impossible to achieve with actual human communications.

The advent of large language models such as ChatGPT suggests a different way forward. While it might be impossible to guarantee security for text created by humans, a new proof lays out for the first time how to achieve perfect security for steganography in machine-generated messages — whether they’re text, images, video or any other media. The authors also include a set of algorithms to produce secure messages, and they are working on ways to combine them with popular apps.

“As we increasingly become a society where it’s very common to interface with AI models, there are increasingly many opportunities to encode secret information in media that people use all the time,” said Samuel Sokota, a computer scientist at Carnegie Mellon University who helped develop the new algorithms.

The result comes from the world of information theory, which provides a mathematical framework for understanding communication of all sorts. It’s an abstract and tidy field, in contrast to the complicated messiness of practical steganography. The worlds don’t often overlap, said Jessica Fridrich, a researcher at Binghamton University who studies ways to hide (and detect) data in digital media. But the new algorithms bring them together by satisfying long-standing theoretical criteria for security and suggesting practical applications for hiding messages in machine-generated content. The new algorithms could be harnessed by spies like the New York Russians, but they could also help people trying to get information in or out of countries that prohibit encrypted channels.

Shaved Heads and Other Strategies

The schemes of steganography, Greek for “covered writing,” predate digital media by millennia.

The earliest known examples show up in The Histories by Herodotus, written in the 5th century BCE. In one story, a message is written on wooden tablets and hidden by a layer of wax to avoid interception during its journey. In another, attributed to Aeneas the Tactician, a message hides dots of invisible ink over certain letters, which spell out the true message. In a more extreme example, the tyrannical leader Histiaeus wants to communicate a strategy to his nephew without detection, so he shaves the head of a slave, tattoos his message on the man’s head and waits for the hair to grow back before sending the messenger. Upon arrival, the nephew shaves the messenger’s head, revealing the plans.

These strategies have persisted, and technology has allowed for new ones. German spies during World War I found ways to transmit information via microdot: They copied and reduced a document until it was as small as the dot of an “i,” which appeared innocent but could be revealed through magnification.

Politicians, too, have turned to the deceptive art. In the 1980s, after a series of press leaks, the British prime minister Margaret Thatcher allegedly had the word processors of her ministers reprogrammed so that each had its own, nigh-undetectable but unique pattern of word spacing. That slight modification allowed leaked documents to be traced to the source.

The approach continues to flourish in the 21st century, for good and evil. Modern steganographic strategies include writing messages in invisible ink (another tactic used by the Russian spies in New York), concealing artist signatures in painting details, and designing audio files with a hidden or backward track. Fridrich says steganographic approaches in digital media can also help hide images in voicemail files or, as in the case of the Russian spies, place written text in doctored photographs.

Formalizing Secrecy

It wasn’t until the 1980s that mathematicians and computer scientists began to seek formal, mathematical rules for steganography, Cachin said. They turned to information theory, a field that had begun with Claude Shannon’s seminal 1948 paper “A Mathematical Theory of Communication,” which established an analytical approach to thinking about sending and receiving information through a channel. (Shannon modeled telegraph lines, but he laid the groundwork for today’s digital technologies.) He used the term “entropy” to quantify the amount of information in a variable — the number of bits required to encode a letter or message, for example — and in 1949 he hammered out rules for perfectly secure cryptography. But Shannon didn’t address security in steganography.

Almost 50 years later, Cachin did. His approach, in the spirit of Shannon, was to think about language probabilistically. Consider two agents, Alice and Bob, who want to communicate a message via steganography and keep it secret from Eve, their adversary. When Alice sends an innocuous message to Bob, she selects words from the entire English lexicon. Those words have probabilities associated with them; for example, the word “the” is more likely to be chosen than, say, “lexicon.” Altogether, the words can be represented as a probability distribution. If Alice uses steganography to send an encoded message to Bob, that message will have its own probability distribution.

Information theorists use a measure called relative entropy to compare probability distributions. It’s like measuring an abstract kind of distance: If the relative entropy between two distributions is zero, “you cannot rely on statistical analysis” to uncover the secret, said Christian Schroeder de Witt, a computer scientist at the University of Oxford who worked on the new paper. In other words, if future spies develop a perfectly secure algorithm to smuggle secrets, no statistics-based surveillance will be able to detect it. Their transmissions will be perfectly hidden.

But Cachin’s proof depended on a critical assumption about the message hiding the secret, known as the cover text. In order to come up with a new message indistinguishable from the original, innocuous one, you have to create a perfect simulation of the cover text distribution, Cachin said. In a written message, for example, that means using some tool that can perfectly simulate a person’s language. But human-generated text is just too messy. It’s possible to come close — ChatGPT and other large language models can produce convincing simulations — but they’re not exact. “For human-generated text, this is not feasible,” Cachin said. For that reason, perfectly secure steganography has long seemed out of reach.

Fridrich, whose research focuses on the complicated real-world intricacies of hiding messages in human-made digital media like photographs and text messages, said perfect simulation is a condition that will never be met. “The problem with digital media is that you will never have that real model,” she said. “It’s too complex. Steganography can never be perfect.”

Achieving Perfection

But machine-generated text, of course, is not created by humans. The recent rise of generative models that focus on language, or others that produce images or sounds, suggests that perfectly secure steganography might be possible in the real world. Those models, after all, use well-defined sampling mechanisms as part of generating text that, in many cases, seems convincingly human.

Sokota and Schroeder de Witt had previously been working not on steganography, but on machine learning. They’d been pursuing new ways to transmit information through various channels, and at one point they learned of a relatively new concept in information theory called a minimum entropy coupling.

“It’s this kind of seemingly fundamental tool that’s not very well explored,” Sokota said. In a minimum entropy coupling, researchers can combine two probability distributions into a single, joint distribution that represents both systems. In the case of steganography, one of those distributions represents the cover text, and the other represents the ciphertext, which contains the hidden message. The joint distribution can ensure that the two texts are statistically indistinguishable, generating a perfectly secure message.

Sokota, Schroeder de Witt and their team had been trying to find ways to exploit the tool for new approaches to deep learning. But one day, Sokota recalled, their collaborator Martin Strohmeier mentioned that their work on minimum entropy coupling reminded him of the security issues around steganography.

Strohmeier was making a casual comment, but Sokota and Schroeder de Witt took it seriously. The group soon figured out how to use a minimum entropy coupling to design a steganographic procedure that met Cachin’s requirements for perfect security in the context of real-world machine learning systems.

“I was surprised to see that it has such a nice application in steganography,” said Murat Kocaoglu, an electrical and computer engineer at Purdue University. He doesn’t work with steganography, but he did help design one of the algorithms the team used in the paper. “This work really ties nicely back to minimum entropy coupling.”

Then the team went further, showing that for a steganography scheme to be as computationally efficient as possible, it must be based on a minimum entropy coupling. The new strategy lays out clear directions for how to achieve both security and efficiency — and suggests that the two go hand in hand.

“Our results seem to suggest that this is even more efficient than approaches that are not perfectly secure,” Sokota said.

The Real World

There are limitations. Cachin pointed out that finding the true minimum entropy coupling is an NP-hard problem, which basically means that the perfect solution is too computationally expensive to be practical, getting back to that issue of efficiency.

Sokota and Schroeder de Witt acknowledge that problem: The optimal coupling would, indeed, be too complicated to compute. But to get around that bottleneck, the authors used an approximating procedure developed by Sokota and Schroeder de Witt (and based on a method introduced by Kocaoglu) that still guarantees security and reasonable efficiency.

Here’s how they see it working in practice: Let’s say that a dissident or a human rights activist wanted to send a text message out of a locked-down country. A plug-in for an app like WhatsApp or Signal would do the heavy algorithmic lifting, Schroeder de Witt said. The first step would be to choose a cover text distribution — that is, a giant collection of possible words to use in the message, as would come from ChatGPT or a similar large language model — that would hide the ciphertext. Then, the program would use that language model to approximate a minimum entropy coupling between the cover text and the ciphertext, and that coupling would generate the string of characters that would be sent by text. To an outside adversary, the new text would be indistinguishable from an innocent machine-generated message. It also wouldn’t have to be text: The algorithm could work by sampling machine-generated art (instead of ChatGPT) or AI-generated audio for voicemails, for example.

The new algorithms are limited in terms of the size of the secret message: Schroeder de Witt estimates that with today’s technology, their system could conceal an image (or other message) of about 225 kilobytes in about 30 seconds of machine-generated voicemail. But it doesn’t need to be enormous to be successful. That’s enough for a substantial message to get past censors or authorities.

Fridrich said she’s more accustomed to working against the limitations of the real world rather than considering the theory. “It’s interesting to see the other side,” she said. For her, the new work starts to bridge the gap between theoretical proofs and real-world messiness. If people don’t use machine-generated content, then the new scheme won’t guarantee security. But as it becomes more widespread, she said, the potential for perfect security will be stronger.

“Everything depends on what will be typical,” she said. If a machine generates a supply of innocuous images that look natural, and people become accustomed to those, then it will be easy to create a source of images enriched with secret messages. “With generative models, this approach gives a possible pathway for the two approaches to meet,” she said.

Clearly, it’s also a double-edged sword. “Criminals will be using it,” Fridrich said, “but it can also be used for good.”

Reprinted with permission from Quanta Magazine, an editorially independent publication of the Simons Foundation whose mission is to enhance public understanding of science by covering research developments and trends in mathematics and the physical and life sciences. Read the original article here.

参考译文
人工智能生成的媒体中可能隐藏秘密信息
2010年6月27日,美国联邦调查局(FBI)在纽约市附近逮捕了10名俄罗斯间谍,他们以美国专业人士的身份生活和工作。这起案件揭开了一个错综复杂的假身份和秘密会面系统,暴露出冷战结束后美国境内最大的间谍网络之一,并启发了电视剧《美国谍梦》的创作。此案也引起人们对隐写术的关注,这是一种将秘密信息伪装在另一种信息中的方式。纽约的间谍们通过公开网站上看似无害的图片中隐藏秘密信息,将通信编码在图像像素中。要解读这些信息,接收者需要下载图片,将其转化为二进制代码中的1和0,并知道哪些被修改过的数字按顺序排列能拼出秘密信息。隐写术是一门既讲究艺术又讲究科学的学科,与更广为人知的密文通信方法——密码学不同。密码学通过将信息内容隐藏、转化为一团文本或数字来实现秘密通信,而隐写术则隐藏了秘密本身的存在。“隐写术隐藏的是信息的存在,”伯尔尼大学的计算机科学家和密码学家Christian Cachin说,“如果对手能够探测到隐含的信息,那么发送者就输掉了这场游戏。”与任何隐蔽通信方式一样,挑战在于如何实现绝对的安全性,即无论是人类还是机器检测器都无法察觉信息中隐藏了秘密。长期以来,隐写术的绝对安全性只是一种理论上的可能,被认为在实际的人类通信中是不可能实现的。随着类似ChatGPT这样的大型语言模型的出现,这一情况似乎可以改写。尽管人类生成的文本不可能保证绝对安全,但一项新的证明首次提出了一种实现机器生成信息中隐写术绝对安全的方法——无论这些信息是文本、图像、视频还是其他媒体。该研究还提供了一组生成安全信息的算法,并正在努力将其与流行的应用程序相结合。“随着我们越来越成为一个与人工智能模型交互成为常态的社会,人们日常使用的媒体中也越来越多地提供了隐藏秘密信息的机会,”卡内基梅隆大学的计算机科学家Samuel Sokota说,他是这些新算法的开发者之一。这一成果源于信息理论的世界,它为理解各类通信提供了数学框架。这是一门抽象且整洁的领域,与实践中的复杂隐写术形成鲜明对比。宾汉顿大学的研究人员Jessica Fridrich说,这两个世界很少重叠。但是,这些新算法通过满足长期存在的安全理论标准,并提出将信息隐藏在机器生成内容中的实用方法,将它们结合了起来。这些新算法可能被像纽约的俄罗斯间谍那样的间谍利用,但也可能帮助那些试图将信息带入或带出禁止加密渠道国家的人们。**剃光头和其他策略**隐写术(希腊语中意为“隐蔽书写”)的历史可追溯到数字媒体出现数千年前。最早的例子出现在公元前5世纪希罗多德的《历史》中。其中一个故事描述了一个信息被写在木板上,然后用一层蜡隐藏,以防止在传递过程中被拦截。另一个故事归功于战术家埃涅阿斯,他通过在某些字母上点上隐形墨水来隐藏信息,从而拼出真实的信息。还有一个更为极端的例子是,暴君希斯提亚斯想与他的侄子秘密沟通策略,于是剃光了奴隶的头,在他的头上纹上信息,等头发长出来后再派他送信。到达后,他的侄子剃掉送信人的头,从而揭示了计划。这些策略一直延续至今,技术的进步也让新的策略不断出现。第一次世界大战期间,德国间谍找到了通过微点发送信息的方法:他们将文件复制并缩小,直到它变得像字母“i”上的点一样大小,看似无害,但可以通过放大发现。政治家们也使用过这种欺骗手法。20世纪80年代,在一系列媒体泄密事件之后,英国首相玛格丽特·撒切尔据称重新编程了她的大臣们的文字处理系统,使每台电脑都有自己几乎无法察觉但独特的单词间距模式。这种微小的修改使泄露的文件可以追踪到源头。进入21世纪后,这种方法依然盛行,既有正面也有负面。现代隐写术的策略包括用隐形墨水书写信息(这也是纽约俄罗斯间谍使用的策略之一)、在绘画细节中隐藏艺术家签名,以及设计包含隐藏或倒放音轨的音频文件。Fridrich表示,数字媒体中的隐写术也可以在机器生成内容中实现。“有了生成模型,这种方法为理论和现实的结合提供了一条可能的路径。”她说。**现实世界中的挑战与前景**尽管这项技术前景广阔,但也存在一些限制。Cachin指出,找到真正的最小熵耦合是一个NP难问题,这意味着完美解决方案的计算成本太高,难以在现实中实现,这又回到了效率的问题。Sokota和Schroeder de Witt承认了这一点:最优耦合确实太复杂,难以计算。但为了绕过这个瓶颈,作者使用了一种由Sokota和Schroeder de Witt开发的近似方法(基于Kocaoglu提出的方法),这种方法仍然能保证安全性和合理效率。以下是他们设想的实际应用方式:假设一名异议人士或人权活动家希望从封锁严重的国家发送一条信息。WhatsApp或Signal等应用程序的插件将承担繁重的算法工作。Schroeder de Witt表示,第一步是选择一个覆盖文本分布——即从ChatGPT或类似的大语言模型中选择一个巨大的可能词汇集——以隐藏密文。然后,程序将使用该语言模型来近似覆盖文本和密文之间的最小熵耦合,并通过该耦合生成将被发送的字符字符串。对于外部对手来说,新文本将与无害的机器生成消息无法区分。它也不一定是文本:算法可以通过采样机器生成的艺术品(而不是ChatGPT)或AI生成的语音邮件音频来实现。新算法在隐藏消息的大小方面有所限制:Schroeder de Witt估计,以目前的技术,他们的系统可以在大约30秒的机器生成语音邮件中隐藏大约225KB的图像(或其他信息)。但这并不需要非常庞大才能成功。这足以让信息通过审查或当局的监管。Fridrich表示,她习惯于在现实世界的限制中工作,而不是考虑理论。“看到另一面很有趣,”她说。对她来说,这项新工作开始弥合理论证明和现实复杂性之间的差距。如果人们不使用机器生成的内容,那么新方案将无法保证安全。但随着其普及度的提高,她说,完美安全的可能性将变得更强。“一切取决于什么将成为典型情况,”她说。如果机器生成的一组看似自然的图像被广泛使用,那么人们习惯了这种图像后,就可以轻松创建一个富含秘密信息的图像来源。“有了生成模型,这种方法为理论和实践的结合提供了一条可能的路径,”她说。“显然,这也是一把双刃剑。”Fridrich说,“罪犯会使用它,但它也可以用于善举。”本文章经Quanta Magazine授权转载。Quanta Magazine是西蒙斯基金会的出版物,其使命是通过报道数学及物理和生命科学领域的研究进展和趋势,增强公众对科学的理解。阅读原始文章请访问此处。
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

广告

scientific

这家伙很懒,什么描述也没留下

关注

点击进入下一篇

ChatGPT能取代心理治疗师吗?

提取码
复制提取码
点击跳转至百度网盘