Quantifying and Analyzing
Entity-level Memorization in Large Language Models
Abstract
Large language models (LLMs) have been proven capable of memorizing their training data, which can be extracted through specifically designed prompts. As the scale of datasets continues to grow, privacy risks arising from memorization have attracted increasing attention. Quantifying language model memorization helps evaluate potential privacy risks. However, prior works on quantifying memorization require access to the precise original data or incur substantial computational overhead, making it difficult for applications in real-world language models. To this end, we propose a fine-grained, entity-level definition to quantify memorization with conditions and metrics closer to real-world scenarios. In addition, we also present an approach for efficiently extracting sensitive entities from autoregressive language models. We conduct extensive experiments based on the proposed, probing language models’ ability to reconstruct sensitive entities under different settings. We find that language models have strong memorization at the entity level and are able to reproduce the training data even with partial leakages. The results demonstrate that LLMs not only memorize their training data but also understand associations between entities. These findings necessitate that trainers of LLMs exercise greater prudence regarding model memorization, adopting memorization mitigation techniques to preclude privacy violations.
Introduction
Pretrained large language models (LLMs) have made significant breakthroughs in various downstream tasks (Devlin et al. 2019; Raffel et al. 2020; Ouyang et al. 2022; OpenAI 2023). However, researches show that LLMs memorize the training data (Carlini et al. 2019, 2021), typically sourced from crowd-sourced corpora or the Internet. Through crafted prompts, language models can reproduce the training data (Huang, Shao, and Chang 2022; Lukas et al. 2023), leading to serious concerns regarding data privacy. As shown in Figure 1, attackers have bypassed the security constraints of language models via the “Grandma Exploit” and successfully extracted sensitive information such as Windows Keys or Apple IDs from ChatGPT. This suggests that privacy leakages in LLMs may be incurred in real-world settings due to emitting memorized sensitive entities. Quantifying the memorization in language models and analyzing their capability of emitting memorized sensitive entities helps researchers understand the potential privacy risks with LLMs.
Recently, several studies have been conducted to quantify memorization in language models. Prior attempts compared models trained under different conditions to evaluate language model memorization (Zhang et al. 2021; Mireshghallah et al. 2022), but these approaches incur high computational costs. Carlini et al. (Carlini et al. 2023) used prefix prompts to prompt models to complete training set suffixes verbatim and quantified memorization based on completion accuracy. However, in real-world scenarios, malicious or curious users are unlikely to have direct access to the training set for obtaining the original prefixes. Additionally, the information leakage from language models is usually fine-grained. For instance, in the “Grandma Exploit” of ChatGPT, the language model has generated content that poses privacy risks. Among these texts, certain key entities (for instance, the keys shown in the figure) constitute sensitive information rather than all contents or verbatim suffixes. Therefore, previous research may not adequately quantify real-world memorization in LLMs.


Apart from the limitations of quantification methodologies, challenges also arise in approaches for extracting memorized training data from language models. Researchers typically employ original prefixes (Carlini et al. 2023) or carefully crafted prompts (Shao et al. 2023; Li et al. 2023) as input. However, obtaining original prefixes is difficult in practical applications of language models. If we forgo the use of prefixes in favor of handcrafted prompts, one must consider that the structure and order (Shin et al. 2020; Lu et al. 2022; Gao, Fisch, and Chen 2021; Jiang et al. 2020) of designed textual prompts can significantly influence the results. In summary, these challenges make it tricky to probe memorization across language model families using textual prompts.
In this paper, we conduct comprehensive testing and evaluation of real-world language models to quantify their memorization. To quantify fine-grained model memorization in a manner that more closely resembles real-world privacy leakages, we propose a definition for memorization extraction at the entity level. We also introduce an approach for learning prompts adaptively, which utilizes entity attribute information and soft prompts (Li and Liang 2021; Lester, Al-Rfou, and Constant 2021). This approach enables efficient language model memorization emitting. Through these, researchers can quantify and analyze LLMs’ memorization under conditions closer to real-world applications. Besides, activating more memorization allows researchers to ascertain potential hazards arising from language models’ memorization. For entity-level specific memorization extraction, our method can achieve the highest accuracy of 61.2%. In terms of average accuracy, our method has a three-fold increase in entity extraction rate in a 6B language model compared to textual prompts when the data is unique to the dataset.
To summarize, our contributions are as follows:
-
1.
We propose a quantifiable, fine-grained definition for evaluating language model memorization without requirements for updating original model parameters or precise data from the training set.
-
2.
We present a method to test language model memorization under conditions that closely resemble real-world scenarios. More memorized information can be extracted by leveraging entity attributes and soft prompts to help understand potential privacy risks.
-
3.
We conduct comprehensive experiments on language model entity-level memorization. Our analysis delves into the processes involved in activating entity memorization in language models and discusses factors that may influence entity memorization.
Related Work
Prompt for LLMs
Pre-trained language models store a vast amount of language knowledge. As models continue to scale up in parameters, methods based on fine-tuning incur high computational overhead. In contrast, prompt-based methods run directly on large language models, achieving results comparable to fine-tuning. Brown et al. (Brown et al. 2020) initially attempted to leverage the language knowledge of the pre-trained language model through the prompts. Shin et al. (Shin et al. 2020) introduced the approach of searching for the optimal prompt word. The prompt generated through this method enhances the utility. Recognizing that prompts’ primary role is to improve models’ performance on specific tasks, researchers have begun exploring approaches that are not constrained by natural language and use continuous vectors directly in the embedding space of the model to prompt language models. This prompt-based approach has yielded many influential extensions, such as p-tuning (Liu et al. 2021b), prompt-tuning (Lester, Al-Rfou, and Constant 2021), prefix-tuning (Li and Liang 2021), etc. These methods have proposed effective solutions for improving model performance.
Privacy Leakage in LLMs
Language models rapidly develop and demonstrate powerful capabilities across various fields (OpenAI 2023). However, language models have some privacy risks associated with data leakage (Béguelin et al. 2020; Mireshghallah et al. 2022). It has recently been confirmed that language models tend to memorize data from their training sets, and prior studies (Huang, Shao, and Chang 2022; Lukas et al. 2023) successfully extracted Personally Identifiable Information (PII) of specific individuals from LLMs.
Previous research (Li et al. 2023; Shao et al. 2023) has thoroughly explored how well-designed prompts may cause LLMs to output data from their training set or related information attributed to their memorization, potentially leading to the leakage of sensitive information. Different language prompts can significantly impact the ultimate predictions (Gao, Fisch, and Chen 2021; Lu et al. 2022). Besides, due to the poor interpretability of LLMs, even experienced researchers struggle to determine how to write better prompts to extract specific PII or entities. Ozdayi et al. (Ozdayi et al. 2023) discussed controllable generation via soft prompts, yet precise prefixes remain necessary. While these methods demonstrate commendable quantification capabilities, they have a challenge in adapting to real-world scenarios, which to some extent, hampers the research into memorization.
Entity Memorization Can Be Extracted Efficiently

Entity Memorization
In fact, for text data, overstrict definitions may overlook some memorized content when key and sensitive information resides only in parts of the data, even though such content is more likely to leak private details. Thus, if model outputs contain partial information from the training data, this can be considered as an emission of memorization. To further elaborate, if the prompt is explicitly targeted, and the portion of the memorized content emitted is sensitive, this constitutes a privacy leak in the language model. In practical applications, private information typically attaches to entities like PII. Since prefix prompts and masked training data rely heavily on original data and are challenging to acquire in practice, obtaining partial entities becomes more accessible than the former two methods. Therefore, we expand the definition of language model memorization to the entity level. We propose a more flexible, quantifiable definition of memorization. When the input prompt is constructed from partial entities in the training strings, and the model’s output contains the expected entity information from the prompt, this is considered as entity memorization.
Definition 1(Entity Memorization).
A training string contains entities and can be uniquely identified by any entities of them. . A prompt contains the n entities that could identify , and expects an entity . Prompted the language model with . If the expected entity E is a substring in the output , the process is referred to as Entity Memorization.
According to Definition 3.1, language models generate the expected entity based on a prompt constructed using the unique identifying entity set from a training string. This quantifies the model’s memorization of training data and its capability for directed reconstruction from memorization. In scenarios of partial training data leakage or malicious querying with extensive guessing, constructing prompt queries from partial entities for the language model to recall unknown entities is more practical. Entity memorization more accurately mirrors real-world memorization leakages by language models than verbatim memorization does. Furthermore, compared to the original prefix or masked training data that strictly emphasize sequential accuracy, entities often have looser relationships in training data. Consequently, Entity Memorization can further reflect models’ understanding and association abilities regarding memorized content.

Efficient Memorization Extraction
Currently, researchers typically measure language model memorization design textual prompts as inputs. These prompts test model memorization in an understandable manner for humans. While textual prompts have good explainability, hand-crafted prompts are highly experience-dependent (Liu et al. 2021a). We find that the performance of textual prompts in extracting fine-grained memorization on subtly related entity levels heavily depends on the scale of the language model, leaving significant room for improvement. Additionally, designing prompts individually for multiple models is laborious and cannot guarantee optimality. These challenges obstruct efficient evaluation of universal language model memorization. Even across models within the same family, crafting universal textual prompts for memorization activation is challenging.
Soft prompt extends prompts into continuous space, providing continuous prompt vectors for the input to the language model. It demonstrates similar workflows across diverse models and excellent performance in downstream tasks. Inspired by soft prompt, we explore applying soft prompts to activate memorization in models. These continuous prompts are typically learned through approaches like prompt tuning. Since the optimal memorization extraction prompt is determined by itself, this approach enables efficient memorization activation while maintaining strong generalization across diverse models.
We introduce Efficient Memorization Extraction (EME) combines effective memorization content filtering and memorization activation. Specifically, EME compresses textual content into entity attribute annotations and combines it with soft prompts as inputs. Through this workflow, testing memorization across different language models follows an identical process, enabling the examination of diverse models’ memorization capacities. We explore applying soft prompts for entity-level memorization activation, with experiments validating fine-grained memorized information extraction. Soft prompt encompasses a wide variety of techniques. When without a specific explanation, the default method used in this work is prefix tuning (Li and Liang 2021).
The lower section of Figure 2 illustrates our approach to implementing a universal entity-level evaluation of language model memorization. For documents that experience partial leakage from the language model’s training set, we extract key entities that uniquely identify this document. Following this, we construct preliminary prompts based on the entity attribute tagging of these entities. Finally, soft prompts are incorporated into the textual prompt vector embeddings to obtain complete prompts as inputs for the model.
We construct textual prompts with entities from training data to prompt the model to generate expected entities and use the method of handcrafted textual prompts as a baseline. The results are shown in Figure 3. As models scale up, their ability to associate entities and emit memorization improve, revealing a more pronounced tendency to emit memorization. Nevertheless, By utilizing entity attributes and soft prompts, the language model’s memorization can be more effectively activated at the entity level. Our approach generates markedly superior performance over textual prompts on larger models. For instance, on GPT-Neo 6B, our approach demonstrates a three-fold increase in memorization extraction efficiency compared to textual prompts.
Experiments and Results
Experimental Setup
The GPT-Neo model family (Black et al. 2021) includes a set of causal language models (CLMs) that are trained on “The Pile” datasets (Gao et al. 2020) and available in four sizes, with 125 million, 1.3 billion, 2.7 billion, and 6 billion parameters respectively. In our experimental setup, we use the greedy decoding strategy by default to generate the output with the minimum perplexity (PPL), which is then utilized for evaluating the model’s entity memorization capabilities.
The Enron email dataset (Klimt and Yang 2004), a subset of “The Pile” dataset, encompasses over 500,000 emails from approximately 150 users of the Enron Corporation. Owing to this dataset’s inherent structured data characteristics, it is conducive to extracting entities from the data, facilitating entity-level memorization extraction experimentation.
For data preprocessing, we extract entities from the Enron dataset and tag their attributes. “Date” and “Content” are retained, while “X-From” and “X-To” are relabeled as “Sender” and “Recipient” attributes for that data instance. The attribute labels “X-From” and “X-To” were modified to be more interpretable for models. These four columns uniquely identify each data instance in the Enron dataset. We select the target entity column as “excepted entity” to recover and then construct prompts from other column entities. To ensure the model searches its memorization rather than deducing answers directly from prompts, we exclude data where the constructed prompt contains the target entity.
Prefix Length Impact on Entity Memorization
The prefix length is a crucial parameter of soft prompts. Typically, a longer prefix length boasts a more robust represention and performs better in downstream tasks.
However, in tests of the model’s memorization capabilities, we find that an optimal extraction rate for memorization can be achieved when the prefix length is short. As the prefix length increases, the entity reconstruction rate of the model declines, then rebounds when the prefix length is extremely long. As depicted in Figure 4, all sizes of GPT-Neo models exhibit a trend of declining memorization extraction rate after reaching the optimal prefix length.


We inspect the text generated by the model and find that when the prefix length is short, the diversity of the generated text is better. In contrast, the output is more conservative when the prefix length is long. This may be attributed to the prefix length ranging from 5 to 10, which tends to optimally activate the model’s memorization. While excessively long prefixes, such as 100, compel the model to prioritize learning the relationships between prompted entities rather than activating memorization. Thus, we conjecture that this relates to model memorization inherently residing in the parameter space, rendering shorter soft prompts sufficient to stimulate innate model memorization.
Different Soft Prompts Have Similar Activation Capabilities
Common soft prompt methodologies, such as prompt-tuning (Lester, Al-Rfou, and Constant 2021) and p-tuning (Liu et al. 2021b), are among those employed in the field. This study also tested the memorization activation capabilities of prompt-tuning and p-tuning.
The results in Figure 5(a) indicate that soft prompt methods can somewhat activate the model’s memorization at the entity level. However, these methods differ in terms of the optimal prefix length setting, suggesting that no universal setting can maximize the memorization of the same model. Furthermore, these methods have different effectiveness in activating the model’s memorization, in our experiments, the prefix tuning approach achieve an entity extraction rate of 9.04%, surpassing the other two methods by a margin of 2%. We speculate that this discrepancy is primarily due to the fact that the prefix-tuning method concatenates the soft prompt at the beginning of the input, which aligns well with the generation characteristics of autoregressive language models.
In addition, we attempt to integrate the results of these methods. In short, we first extract entities from the generated text of different methods. If the same output entity is present, we select that entity as the result. Otherwise, we embed the expected entity from each method back into the input tensor and let the model evaluate the perplexity of the input. We choose the method with the lowest perplexity as the final output result. The results show that this method can, to some extent, integrate the advantages of these methods under different settings and achieve better robustness. However, it does not show any significant advantage in accurcy of memorization reconstruction, with a prefix length of 5, the activate ensemble only achieve a improvement of 0.4%.
Model | Fabricated-50 | Fabricated-100 | Fabricated-200 | Fabricated-500 | Fabricated-1000 |
GPT-Neo 125M | 0.0 ± 0 | 0.0 ± 0 | 0.0 ± 0 | 0.0 ± 0 | 0.0 ± 0 |
GPT-Neo 1.3B | 0.0 ± 0.003 | 0.0 ± 0.002 | 0.0 ± 0.002 | 0.001 ± 0.002 | 0.001 ± 0.001 |
GPT-Neo 2.7B | 0.004 ± 0.008 | 0.006 ± 0.010 | 0.007 ± 0.006 | 0.007 ± 0.003 | 0.006 ± 0.003 |
GPT-J 6B | 0.019 ± 0.016 | 0.036 ± 0.019 | 0.057 ± 0.012 | 0.057 ± 0.009 | 0.052 ± 0.014 |
OPT 6.7B | 0.0 ± 0 | 0.0 ± 0 | 0.0 ± 0 | 0.0 ± 0.001 | 0.0 ± 0.001 |
LLAMA 7B | 0.0 ± 0 | 0.0 ± 0 | 0.0 ± 0.001 | 0.0 ± 0.002 | 0.001 ± 0.003 |
Real-50 | Real-100 | Real-200 | Real-500 | Real-1000 | |
---|---|---|---|---|---|
GPT-Neo 125M | 0.003 ± 0.002 | 0.008 ± 0.003 | 0.009 ± 0.003 | 0.008 ± 0.004 | 0.007 ± 0.006 |
GPT-Neo 1.3B | 0.009 ± 0.002 | 0.015 ± 0.004 | 0.019 ± 0.003 | 0.016 ± 0.004 | 0.016 ± 0.009 |
GPT-Neo 2.7B | 0.014 ± 0.003 | 0.019 ± 0.005 | 0.020 ± 0.006 | 0.020 ± 0.009 | 0.019 ± 0.008 |
GPT-J 6B | 0.052 ± 0.009 | 0.072 ± 0.013 | 0.090 ± 0.016 | 0.090 ± 0.013 | 0.088 ± 0.021 |
OPT 6.7B | 0.0 ± 0.001 | 0.001 ± 0.002 | 0.002 ± 0.001 | 0.001 ± 0.002 | 0.002 ± 0.002 |
LLAMA 7B | 0.001 ± 0.001 | 0.001 ± 0.001 | 0.003 ± 0.002 | 0.004 ± 0.003 | 0.002 ± 0.003 |
LLMs Struggle to Reconstruct Entities without Memorization
In an effort to specifically exclude the possibility of language models learning entity relationships through soft prompts, we conduct an ablation study on LLM proposed by Meta. OPT (Zhang et al. 2022) utilizes a portion of “The Pile” datasets but does not include the Enron dataset. Similarly, Another large-scale model trained by Meta, Llama (Touvron et al. 2023), also does not explicitly mention the usage of the Enron dataset. We conduct tests with the same hyperparameter settings on the OPT-6.7B and LLAMA-7B models, similar in size to the GPT-J 6B model.
The experimental results shown in Figure 5(b) indicate that OPT and Llama, lacking training data memorization, achieve a maximum accuracy of 0.4% in accurately reconstructing the expected entities when utilizing soft prompts. The accuracy remains relatively low despite the models’ fitting ability from the prompt improving with increasing prefix length. Their performances are even inferior to the smaller GPT-Neo model, which possesses memorization. Without memorization, LLMs struggle to learn the subtle relationships between entities based on soft prompts and cannot reconstruct the expected entity. Reconstruction can only be completed if LLMs have learned the dataset during pre-training.
Based on the above results, we believe that soft prompts indeed activate the model’s memorization rather than enabling the model to learn a lower-loss generation method.
Fabricated Data Also Activate Memorization
In order to further evaluate the extent to which the model can generate mnemonic content under more lenient conditions, we attempt to construct soft prompts using fabricated data. We randomly recombine the target entities with other remaining entities, ensuring new entity pairs are not in the original dataset’s documents.
Table 1 depicts the accuracy of model reconstruction of correct entities using varying quantities of real and fabricated data. The results show that when training soft prompts with fabricated data and providing prompts to the model, despite a near 4% loss in reconstruction rate compared to real data, the model can still generate the expected entity from memorization with a maximum accuracy of 5.7%. With smaller-sized models, soft prompts obtained from fabricated data are less effective, and the final results fluctuate greatly, but larger-sized models with better robustness are more adaptable. We compare how models without memorization respond to soft prompts constructed by fabricated data, and the results showed that learning even the relationships between entities becomes more challenging on fabricated data for these models.
Therefore, we believe that the language models do not merely reproduce entities superficially based on the data used during soft prompt training. Even if the data is fabricated and contains biases, the soft prompts obtained from the fabricated data can still evoke the model’s memorization of the related dataset to a certain extent. Furthermore, this underscores the potential privacy risks associated with the model’s memorization at the entity level. Attackers might be able to access the model’s memorization under more lenient conditions. Even if they do not have the real dataset, knowing what entities exist in a dataset can allow them to construct prompts to retrieve sensitive information.
Training Data Volume Influence Performance of Soft Prompts
For optimal soft prompt performance, the model must learn from training data. Consequently, the data choice will impact the effectiveness of the soft prompts. We investigat how the amount of data affects the performance of soft prompts in the extraction of entity memorization.
When using real data, we find that soft prompts trained from less data often fail to activate the model’s memorization effectively. This outcome is little influenced by the randomness in choosing training data. As the dataset size increases, the effectiveness of the soft prompts improves and eventually stabilizes. However, the effectiveness declines with massive training datasets, and the results exhibit some fluctuations. Upon examining these training data divisions, we argue that an abundance of training data might lead soft prompts to lose some of their memorization activation effectiveness. They instead turn to learning how to derive correct answers directly from the training data.
The variance in the entity extraction rate is different when using fabricated data than using real data. The results vary more greatly when using a small amount of fabricated data, and we speculate that this variance is primarily related to the randomness during data fabrication. On the other hand, when using more fabricated data, the variance initially diminishes but later rebounds. We investigate the cause of this and contend that it may be related to the data fabrication method we chose. Because our fabricated data do not exist in the original data and entity pairs are not duplicated, this differs from real data and may eventually lead to the variances observed in the results.

Duplicate Times Strengthen Entity Memorization
Considering previous research, a popular view proposed by (Kandpal, Wallace, and Raffel 2022; Lee et al. 2022) is that the memorization of the language model is strongly correlated with the frequency of data duplication. Data that duplicate excessively in the training dataset is more likely to be extracted, yet previous research has not explored this issue on the entity level. Therefore, we statistically analyze the duplications of uniquely identified entities and test the reconstruction rates of entities with different numbers of repetitions.
We bin duplicated entity sequences with similar duplicated counts and use the same settings to obtain soft prompts for input construction. And then query the model multiple times to gauge the effectiveness of the entity sequence in extracting the expected entity. As the results in Figure 6 indicate, even though the entity sequence uniquely identifies a unique document and repeats less frequently in other content, the repetition count of some entity sequences still affects the degree to which the model memorizes them. After the entity sequence is duplicated more than 50 times, the mean accuracy of query results exceeded 10%. Beyond 200 duplications, the reconstruction accuracy of expected entity reached a maximum of 61.2%. In other words, when entity sequences are extensively duplicated, the model has stronger memorization of these data, presenting a higher risk of privacy leakage.
Since the model’s entity-level memorization still correlates with the duplication times, deduplicating data still has a mitigating effect on memorization at the entity level. However, duplication at the entity level is inevitable since these documents will not be remarkably similar in other aspects, which makes it even more challenging to guard against malicious queries through deduplication.
Limitations
Due to privacy concerns, using undisclosed or potentially privacy-infringing datasets is strictly prohibited. Therefore, we performed structured processing and experimentation exclusively on the Enron dataset. Empirical results have demonstrated that language models exhibit sufficiently strong extraction capabilities at the entity level. Consequently, if LLMs training or fine-tuning is performed on similar semi-structured sensitive data, the entity extraction results will be replicable.
Structured or semi-structured data with abundant entity information is relatively rare among common LLMs training datasets, posing challenges for finding multiple sets to experiment on extracting unfamiliar sensitive entities. The necessity of procuring a large amount of structured data for testing purposes further exacerbates this problem. Therefore, despite our methodology being applicable across various models, we have been unable to conduct cross-family model experiments due to these constraints. Language model memorization markedly strengthens with increasing model size within the same family, but large model inference has high computational demands. Due to computational constraints, we only conducted experiments on the GPT-Neo family, with the largest LLMs scale being 6B.
Conclusions and Future Work
In our paper, we introduce a definition of entity-level memorization for language models to quantify fine-grained memorization. With this, researchers can effectively quantify language model memorization to reproduce training data that mirrors real-world scenarios. Our method merges entity attribute information with trainable soft prompts to prompt the language models. Notably, the results indicate that our method can significantly activate memorization with subtle entity associations in the model, even if the data appears only once in the dataset. The entity-level memorization of specific individual data is more explicit than the verbatim memorization previously proposed, making entity-level memorization better quantify privacy risks.
We comprehensively analyze entity-level memorization of the language models by conducting experiments. Our results show that language models have memorized an amount of their training data. Through appropriate prompts, LLMs can generate fine-grained sensitive information with a high probability. This reminds that researchers of LLMs need to pay serious attention to privacy issues in this field, especially language models trained on sensitive datasets. It is imperative to employ strategies that either mitigate memorization or establish defenses to minimize associated risks.
In future work, we intend to investigate more effective extraction strategies to evaluate the memorization of language models and scale up to larger models to explore their entity-level memorization. Furthermore, we also aim to develop methods to mitigate entity-level memorization.
References
- Béguelin et al. (2020) Béguelin, S. Z.; Wutschitz, L.; Tople, S.; Rühle, V.; Paverd, A.; Ohrimenko, O.; Köpf, B.; and Brockschmidt, M. 2020. Analyzing Information Leakage of Updates to Natural Language Models. In Ligatti, J.; Ou, X.; Katz, J.; and Vigna, G., eds., CCS ’20: 2020 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, USA, November 9-13, 2020, 363–375. ACM.
- Black et al. (2021) Black, S.; Leo, G.; Wang, P.; Leahy, C.; and Biderman, S. 2021. GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow. If you use this software, please cite it using these metadata.
- Brown et al. (2020) Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J. D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; Agarwal, S.; Herbert-Voss, A.; Krueger, G.; Henighan, T.; Child, R.; Ramesh, A.; Ziegler, D.; Wu, J.; Winter, C.; Hesse, C.; Chen, M.; Sigler, E.; Litwin, M.; Gray, S.; Chess, B.; Clark, J.; Berner, C.; McCandlish, S.; Radford, A.; Sutskever, I.; and Amodei, D. 2020. Language Models are Few-Shot Learners. In Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M.; and Lin, H., eds., Advances in Neural Information Processing Systems, volume 33, 1877–1901. Curran Associates, Inc.
- Carlini et al. (2023) Carlini, N.; Ippolito, D.; Jagielski, M.; Lee, K.; Tramer, F.; and Zhang, C. 2023. Quantifying Memorization Across Neural Language Models. In The Eleventh International Conference on Learning Representations.
- Carlini et al. (2019) Carlini, N.; Liu, C.; Erlingsson, Ú.; Kos, J.; and Song, D. 2019. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks. In Heninger, N.; and Traynor, P., eds., 28th USENIX Security Symposium, USENIX Security 2019, Santa Clara, CA, USA, August 14-16, 2019, 267–284. USENIX Association.
- Carlini et al. (2021) Carlini, N.; Tramèr, F.; Wallace, E.; Jagielski, M.; Herbert-Voss, A.; Lee, K.; Roberts, A.; Brown, T. B.; Song, D.; Erlingsson, Ú.; Oprea, A.; and Raffel, C. 2021. Extracting Training Data from Large Language Models. In Bailey, M.; and Greenstadt, R., eds., 30th USENIX Security Symposium, USENIX Security 2021, August 11-13, 2021, 2633–2650. USENIX Association.
- Devlin et al. (2019) Devlin, J.; Chang, M.; Lee, K.; and Toutanova, K. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Burstein, J.; Doran, C.; and Solorio, T., eds., Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers), 4171–4186. Association for Computational Linguistics.
- Gao et al. (2020) Gao, L.; Biderman, S.; Black, S.; Golding, L.; Hoppe, T.; Foster, C.; Phang, J.; He, H.; Thite, A.; Nabeshima, N.; Presser, S.; and Leahy, C. 2020. The Pile: An 800GB Dataset of Diverse Text for Language Modeling. arXiv:2101.00027.
- Gao, Fisch, and Chen (2021) Gao, T.; Fisch, A.; and Chen, D. 2021. Making Pre-trained Language Models Better Few-shot Learners. In Zong, C.; Xia, F.; Li, W.; and Navigli, R., eds., Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, 3816–3830. Association for Computational Linguistics.
- Huang, Shao, and Chang (2022) Huang, J.; Shao, H.; and Chang, K. C. 2022. Are Large Pre-Trained Language Models Leaking Your Personal Information? In Goldberg, Y.; Kozareva, Z.; and Zhang, Y., eds., Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, 2038–2047. Association for Computational Linguistics.
- Jiang et al. (2020) Jiang, Z.; Xu, F. F.; Araki, J.; and Neubig, G. 2020. How Can We Know What Language Models Know. Trans. Assoc. Comput. Linguistics, 8: 423–438.
- Kandpal, Wallace, and Raffel (2022) Kandpal, N.; Wallace, E.; and Raffel, C. 2022. Deduplicating Training Data Mitigates Privacy Risks in Language Models. In Chaudhuri, K.; Jegelka, S.; Song, L.; Szepesvári, C.; Niu, G.; and Sabato, S., eds., International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, 10697–10707. PMLR.
- Klimt and Yang (2004) Klimt, B.; and Yang, Y. 2004. The Enron Corpus: A New Dataset for Email Classification Research. In Boulicaut, J.; Esposito, F.; Giannotti, F.; and Pedreschi, D., eds., Machine Learning: ECML 2004, 15th European Conference on Machine Learning, Pisa, Italy, September 20-24, 2004, Proceedings, volume 3201 of Lecture Notes in Computer Science, 217–226. Springer.
- Lee et al. (2022) Lee, K.; Ippolito, D.; Nystrom, A.; Zhang, C.; Eck, D.; Callison-Burch, C.; and Carlini, N. 2022. Deduplicating Training Data Makes Language Models Better. In Muresan, S.; Nakov, P.; and Villavicencio, A., eds., Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, 8424–8445. Association for Computational Linguistics.
- Lester, Al-Rfou, and Constant (2021) Lester, B.; Al-Rfou, R.; and Constant, N. 2021. The Power of Scale for Parameter-Efficient Prompt Tuning. In Moens, M.; Huang, X.; Specia, L.; and Yih, S. W., eds., Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, 3045–3059. Association for Computational Linguistics.
- Li et al. (2023) Li, H.; Guo, D.; Fan, W.; Xu, M.; Huang, J.; Meng, F.; and Song, Y. 2023. Multi-step Jailbreaking Privacy Attacks on ChatGPT. arXiv:2304.05197.
- Li and Liang (2021) Li, X. L.; and Liang, P. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In Zong, C.; Xia, F.; Li, W.; and Navigli, R., eds., Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, ACL/IJCNLP 2021, (Volume 1: Long Papers), Virtual Event, August 1-6, 2021, 4582–4597. Association for Computational Linguistics.
- Liu et al. (2021a) Liu, P.; Yuan, W.; Fu, J.; Jiang, Z.; Hayashi, H.; and Neubig, G. 2021a. Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing. CoRR, abs/2107.13586.
- Liu et al. (2021b) Liu, X.; Zheng, Y.; Du, Z.; Ding, M.; Qian, Y.; Yang, Z.; and Tang, J. 2021b. GPT Understands, Too. arXiv:2103.10385.
- Lu et al. (2022) Lu, Y.; Bartolo, M.; Moore, A.; Riedel, S.; and Stenetorp, P. 2022. Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity. In Muresan, S.; Nakov, P.; and Villavicencio, A., eds., Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, 8086–8098. Association for Computational Linguistics.
- Lukas et al. (2023) Lukas, N.; Salem, A.; Sim, R.; Tople, S.; Wutschitz, L.; and Zanella-Béguelin, S. 2023. Analyzing Leakage of Personally Identifiable Information in Language Models. In 2023 2023 IEEE Symposium on Security and Privacy (SP) (SP), 346–363. Los Alamitos, CA, USA: IEEE Computer Society.
- Mireshghallah et al. (2022) Mireshghallah, F.; Uniyal, A.; Wang, T.; Evans, D.; and Berg-Kirkpatrick, T. 2022. An Empirical Analysis of Memorization in Fine-tuned Autoregressive Language Models. In Goldberg, Y.; Kozareva, Z.; and Zhang, Y., eds., Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, 1816–1826. Association for Computational Linguistics.
- OpenAI (2023) OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774.
- Ouyang et al. (2022) Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; Schulman, J.; Hilton, J.; Kelton, F.; Miller, L.; Simens, M.; Askell, A.; Welinder, P.; Christiano, P. F.; Leike, J.; and Lowe, R. 2022. Training language models to follow instructions with human feedback. In Koyejo, S.; Mohamed, S.; Agarwal, A.; Belgrave, D.; Cho, K.; and Oh, A., eds., Advances in Neural Information Processing Systems, volume 35, 27730–27744. Curran Associates, Inc.
- Ozdayi et al. (2023) Ozdayi, M.; Peris, C.; FitzGerald, J.; Dupuy, C.; Majmudar, J.; Khan, H.; Parikh, R.; and Gupta, R. 2023. Controlling the Extraction of Memorized Data from Large Language Models via Prompt-Tuning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 1512–1521. Toronto, Canada: Association for Computational Linguistics.
- Raffel et al. (2020) Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; and Liu, P. J. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res., 21: 140:1–140:67.
- Shao et al. (2023) Shao, H.; Huang, J.; Zheng, S.; and Chang, K. C.-C. 2023. Quantifying Association Capabilities of Large Language Models and Its Implications on Privacy Leakage. arXiv:2305.12707.
- Shin et al. (2020) Shin, T.; Razeghi, Y.; IV, R. L. L.; Wallace, E.; and Singh, S. 2020. AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts. In Webber, B.; Cohn, T.; He, Y.; and Liu, Y., eds., Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, 4222–4235. Association for Computational Linguistics.
- Touvron et al. (2023) Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.-A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; Rodriguez, A.; Joulin, A.; Grave, E.; and Lample, G. 2023. LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971.
- Zhang et al. (2021) Zhang, C.; Ippolito, D.; Lee, K.; Jagielski, M.; Tramèr, F.; and Carlini, N. 2021. Counterfactual Memorization in Neural Language Models. arXiv:2112.12938.
- Zhang et al. (2022) Zhang, S.; Roller, S.; Goyal, N.; Artetxe, M.; Chen, M.; Chen, S.; Dewan, C.; Diab, M.; Li, X.; Lin, X. V.; Mihaylov, T.; Ott, M.; Shleifer, S.; Shuster, K.; Simig, D.; Koura, P. S.; Sridhar, A.; Wang, T.; and Zettlemoyer, L. 2022. OPT: Open Pre-trained Transformer Language Models. arXiv:2205.01068.