OpenAI says ChatGPT’s Memory is an opt-in feature from the start, and can be wiped at any point, either in settings or by simply instructing the bot to wipe it. Once the Memory setting is cleared, that information won’t be used to train its AI model. It’s unclear how exactly much of that personal data is used to train the AI while someone is chatting with the chatbot. And toggling off Memory does not mean you’ve totally opted out of having your chats train OpenAI’s model; that’s a separate opt-out.
The company also claims that it won’t store certain sensitive information in Memory. If you tell ChatGPT your password (don’t do this) or Social Security number (or this), the app’s Memory is thankfully forgetful. Jang also says OpenAI is still soliciting feedback on whether other personally identifiable information, like a user’s ethnicity, is too sensitive for the company to auto-capture.
“We think there are a lot of useful cases for that example, but for now we have trained the model to steer away from proactively remembering that information,” Jang says.
It’s easy to see how ChatGPT’s Memory function could go awry, instances where a user might have forgotten they once asked the chatbot about a kink or an abortion clinic or a non-violent way to deal with a mother-in-law, only to be reminded of it—or have others see it—in a future chat. How ChatGPT’s Memory handles health data is also something of an open question. “We steer ChatGPT away from remembering certain health details but this is still a work in progress,” says OpenAI spokesperson Niko Felix. In this way ChatGPT is the same song, just a new era, about the internet’s permanence: Look at this great new Memory feature, until it’s a bug.
OpenAI is also not the first entity to toy with memory in generative AI. Google has emphasized “multi-turn” technology in Gemini 1.0, its own LLM. This means you can interact with Gemini Pro using a single-turn prompt—one back-and-forth between the user and the chatbot—or have a multi-turn, continuous conversation in which the bot “remembers” the context from previous messages.
An AI framework company called LangChain has been developing a Memory module that helps large language models recall previous interactions between an end user and the model. Giving LLMs a long-term memory “can be very powerful in creating unique LLM experiences—a chatbot can begin to tailor its responses towards you as an individual, based on what it knows about you,” says Harrison Chase, cofounder and CEO of LangChain. “The lack of long-term memory can also create a grating experience. No one wants to have to tell a restaurant-recommendation chatbot over and over that they are vegetarian.”
This technology is sometimes referred to as “context retention” or “persistent context” rather than “memory,” but the end goal is the same: for the human-computer interaction to feel so fluid, so natural, that the user can easily forget what the chatbot might remember. This is also a potential boon for businesses deploying these chatbots, who want to maintain an ongoing relationship with the customer on the other end.
“You can think of these as just a number of tokens that are getting prepended to your conversations,” says Liam Fedus, an OpenAI research scientist. “The bot has some intelligence, and behind the scenes it’s looking at the memories and saying, ‘These look like they’re related; let me merge them.’ And that then goes on your token budget.”
Fedus and Jang say that ChatGPT’s memory is nowhere near the capacity of the human brain. And yet, in almost the same breath, Fedus explains that with ChatGPT’s memory, you’re limited to “a few thousand tokens.” If only.
Is this the hyper-vigilant virtual assistant tech consumers have been promised for the past decade, or just another data-capture scheme that uses your likes and preferences and personal data to better serve a tech company than its users? Possibly both, though OpenAI might not put it that way. “I think the assistants of the past just didn’t have the intelligence,” Fedus said, “and now we’re getting there.”
Will Knight contributed to this story.
Source