Shared References

196 / 196 references
LLM Social Simulations Are a Promising Research Method 2025 article

Jacy Reese Anthis, Ryan Liu, Sean M. Richardson, Austin C. Kozlowski, Bernard Koch, James Evans, Erik Brynjolfsson, Michael Bernstein

DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning 2025 article

{DeepSeek AI}

Can We Trust AI Benchmarks? An Interdisciplinary Review of Current Issues in AI Evaluation 2025 article

Maria Eriksson, Erasmo Purificato, Arman Noroozian, Joao Vinagre, Guillaume Chaslot, Emilia Gomez, David Fernandez-Llorca

Exploring the limits of strong membership inference attacks on large language models 2025 article

Jamie Hayes, Ilia Shumailov, Christopher A. Choquette-Choo, Matthew Jagielski, George Kaissis, Milad Nasr, Sahra Ghalebikesabi, Meenatchi Sundaram Mutu Selva Annamalai, Niloofar Mireshghallah, Igor Shilov, Matthieu Meeus, Yves-Alexandre de Montjoye, Katherine Lee, Franziska Boenisch, Adam Dziedzic, A. Feder Cooper

Trust and Friction: Negotiating How Information Flows Through Decentralized Social Media 2025 article

Hwang, Sohyeon, Nanayakkara, Priyanka, Shvartzshnaider, Yan

Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development 2025 article

Jan Kulveit, Raymond Douglas, Nora Ammann, Deger Turan, David Krueger, David Duvenaud

Extending "GPTs Are GPTs" to Firms 2025 article

Benjamin Labaschin, Tyna Eloundou, Sam Manning, Pamela Mishkin, Daniel Rock

OLMoTrace: Tracing Language Model Outputs Back to Trillions of Training Tokens 2025 article

Liu, Jiacheng, Blanton, Taylor, Elazar, Yanai, Min, Sewon, Chen, YenSung, Chheda-Kothary, Arnavi, Tran, Huy, Bischoff, Byron, Marsh, Eric, Schmitz, Michael, others

Cybernetics 2025 misc

Cybernetics is the transdisciplinary study of circular causal processes such as feedback and recursion, where the effects of a system's actions (its outputs) return as inputs to that system, influencing subsequent action. It is concerned with general principles that are relevant across multiple contexts, including in engineering, ecological, economic, biological, cognitive and social systems and also in practical activities such as designing, learning, and managing. Cybernetics' transdisciplinary character has meant that it intersects with a number of other fields, leading to it having both wide influence and diverse interpretations. The field is named after an example of circular causal feedback—that of steering a ship (the ancient Greek κυβερνήτης (kybernḗtēs) refers to the person who steers a ship). In steering a ship, the position of the rudder is adjusted in continual response to the effect it is observed as having, forming a feedback loop through which a steady course can be maintained in a changing environment, responding to disturbances from cross winds and tide. Cybernetics has its origins in exchanges between numerous disciplines during the 1940s. Initial developments were consolidated through meetings such as the Macy Conferences and the Ratio Club. Early focuses included purposeful behaviour, neural networks, heterarchy, information theory, and self-organising systems. As cybernetics developed, it became broader in scope to include work in design, family therapy, management and organisation, pedagogy, sociology, the creative arts and the counterculture.

The Leaderboard Illusion 2025 article

Shivalika Singh, Yiyang Nan, Alex Wang, Daniel D'Souza, Sayash Kapoor, Ahmet Üstün, Sanmi Koyejo, Yuntian Deng, Shayne Longpre, Noah A. Smith, Beyza Ermis, Marzieh Fadaee, Sara Hooker

If open source is to win, it must go public 2025 article

Tan, Joshua, Vincent, Nicholas, Elkins, Katherine, Sahlgren, Magnus

Canada as a Champion for Public AI: Data, Compute and Open Source Infrastructure for Economic Growth and Inclusive Innovation 2025 article

Vincent, Nicholas, Surman, Mark, Hirsch-Allen, Jake

Quantitative Analysis of AI-Generated Texts in Academic Research: A Study of AI Presence in Arxiv Submissions using AI Detection Tool 2024 article

Arslan Akram

The Illusion of Artificial Inclusion 2024 inproceedings

William Agnew, A. Stevie Bergman, Jennifer Chien, Mark Diaz, Seliem El-Sayed, Jaylen Pittman, Shakir Mohamed, Kevin R. McKee

The Multilingual Alignment Prism: Aligning Global and Local Preferences to Reduce Harm 2024 article

Aakanksha, Arash Ahmadian, Beyza Ermis, Seraphina Goldfarb-Tarrant, Julia Kreutzer, Marzieh Fadaee, Sara Hooker

Machines of Loving Grace: How AI Could Transform the World for the Better 2024 misc

Amodei, Dario

To code, or not to code? exploring impact of code in pre-training 2024 article

Aryabumi, Viraat, Su, Yixuan, Ma, Raymond, Morisot, Adrien, Zhang, Ivan, Locatelli, Acyr, Fadaee, Marzieh, {\"U}st{\"u}n, Ahmet, Hooker, Sara

The Rise of AI-Generated Content in Wikipedia 2024 article

Creston Brooks, Samuel Eggert, Denis Peskoff

Poisoning Web-Scale Training Datasets is Practical 2024 misc

Nicholas Carlini, Matthew Jagielski, Christopher A. Choquette-Choo, Daniel Paleka, Will Pearce, Hyrum Anderson, Andreas Terzis, Kurt Thomas, Florian Tramèr

What is Your Data Worth to GPT? LLM-Scale Data Valuation with Influence Functions 2024 article

Choe, Sang Keun and Ahn, Hwijeen and Bae, Juhan and Zhao, Kewen and Kang, Minsoo and Chung, Youngseog and Pratapa, Adithya and Neiswanger, Willie and Strubell, Emma and Mitamura, Teruko and Schneider, Jeff and Hovy, Eduard and Grosse, Roger and Xing, Eric

Large language models reduce public knowledge sharing on online Q&A platforms 2024 article

Del Rio-Chanona, R. Maria and Laurentsyeva, Nadzeya and Wachs, Johannes

Direct Preference Optimization: Your Language Model is Secretly a Reward Model 2024 article

Rafael Rafailov, Archit Sharma, Eric Mitchell, Stefano Ermon, Christopher D. Manning, Chelsea Finn

GPTs are GPTs: Labor Market Impact Potential of LLMs 2024 article

Tyna Eloundou, Sam Manning, Pamela Mishkin, Daniel Rock

Artificial Intelligence Act 2024 misc

{European Union}

Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training 2024 misc

Evan Hubinger, Carson Denison, Jesse Mu, Mike Lambert, Meg Tong, Monte MacDiarmid, Tamera Lanham, Daniel M. Ziegler, Tim Maxwell, Newton Cheng, Adam Jermyn, Amanda Askell, Ansh Radhakrishnan, Cem Anil, David Duvenaud, Deep Ganguli, Fazl Barez, Jack Clark, Kamal Ndousse, Kshitij Sachan, Michael Sellitto, Mrinank Sharma, Nova DasSarma, Roger Grosse, Shauna Kravec, Yuntao Bai, Zachary Witten, Marina Favaro, Jan Brauner, Holden Karnofsky, Paul Christiano, Samuel R. Bowman, Logan Graham, Jared Kaplan, Sören Mindermann, Ryan Greenblatt, Buck Shlegeris, Nicholas Schiefer, Ethan Perez

Public {AI}: {Infrastructure} for the {Common} {Good} 2024 techreport

Jackson, Brandon, Cavello, B, Devine, Flynn, Garcia, Nick, Klein, Samuel J., Krasodomski, Alex, Tan, Joshua, Tursman, Eleanor

ANSI/NISO Z39.96-2024, JATS: Journal Article Tag Suite 2024 techreport
Wikimedia data for AI: a review of Wikimedia datasets for NLP tasks and AI-assisted editing 2024 article

Johnson, Isaac, Kaffee, Lucie-Aim{\'e}e, Redi, Miriam

Data {Flywheel} {Go} {Brrr}: {Using} {Your} {Users} to {Build} {Better} {Products} - {Jason} {Liu} 2024 misc

Liu, Jason

Explore how data flywheels leverage user feedback to enhance product development and achieve business success with AI.

Consent in Crisis: The Rapid Decline of the AI Data Commons 2024 article

Longpre, Shayne and Mahari, Robert and Lee, Ariel and Lund, Campbell and Oderinwale, Hamidah and Brannon, William and Saxena, Nayan and Obeng-Marnu, Naana and Sud, Tobin and Gupta, Sameer and Muennighoff, Niklas and others

StarCoder 2 and The Stack v2: The Next Generation 2024 misc

Anton Lozhkov, Raymond Li, Loubna Ben Allal, Federico Cassano, Joel Lamy-Poirier, Nouamane Tazi, Ao Tang, Dmytro Pykhtar, Jiawei Liu, Yuxiang Wei, Tianyang Liu, Max Tian, Denis Kocetkov, Arthur Zucker, Younes Belkada, Zijian Wang, Qian Liu, Dmitry Abulkhanov, Indraneil Paul, Zhuang Li, Wen-Ding Li, Megan Risdal, Jia Li, Jian Zhu, Terry Yue Zhuo, Evgenii Zheltonozhskii, Nii Osae Osae Dade, Wenhao Yu, Lucas Krauß, Naman Jain, Yixuan Su, Xuanli He, Manan Dey, Edoardo Abati, Yekun Chai, Niklas Muennighoff, Xiangru Tang, Muhtasham Oblokulov, Christopher Akiki, Marc Marone, Chenghao Mou, Mayank Mishra, Alex Gu, Binyuan Hui, Tri Dao, Armel Zebaze, Olivier Dehaene, Nicolas Patry, Canwen Xu, Julian McAuley, Han Hu, Torsten Scholak, Sebastien Paquet, Jennifer Robinson, Carolyn Jane Anderson, Nicolas Chapados, Mostofa Patwary, Nima Tajbakhsh, Yacine Jernite, Carlos Muñoz Ferrandis, Lingming Zhang, Sean Hughes, Thomas Wolf, Arjun Guha, Leandro von Werra, Harm de Vries

LLM Dataset Inference: Did you train on my dataset? 2024 article

Pratyush Maini, Hengrui Jia, Nicolas Papernot, Adam Dziedzic

Public AI: Making AI Work for Everyone, by Everyone 2024 misc

Marda, Nik, Sun, Jasmine, Surman, Mark

Scalable Data Ablation Approximations for Language Models through Modular Training and Merging 2024 inproceedings

Na, Clara and Magnusson, Ian and Jha, Ananya Harsh and Sherborne, Tom and Strubell, Emma and Dodge, Jesse and Dasigi, Pradeep

Generative AI Profile (Draft/2024) 2024 techreport
A Canary in the AI Coal Mine: American Jews May Be Disproportionately Harmed by Intellectual Property Dispossession in Large Language Model Training 2024 article

Heila Precel, Allison McDonald, Brent Hecht, Nicholas Vincent

What is a {Data} {Flywheel}? {A} {Guide} to {Sustainable} {Business} {Growth} 2024 misc

Roche, Adam, Sassoon, Yali

The data addition dilemma 2024 article

Shen, Judy Hanwen, Raji, Inioluwa Deborah, Chen, Irene Y

Data {Flywheels} for {LLM} {Applications} 2024 misc

Shankar, Shreya

Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research 2024 article

Luca Soldaini, Rodney Kinney, Akshita Bhagia, Dustin Schwenk, David Atkinson, Russell Authur, Ben Bogin, Khyathi Chandu, Jennifer Dumas, Yanai Elazar, Valentin Hofmann, Ananya Harsh Jha, Sachin Kumar, Li Lucy, Xinxi Lyu, Nathan Lambert, Ian Magnusson, Jacob Morrison, Niklas Muennighoff, Aakanksha Naik, Crystal Nam, Matthew E. Peters, Abhilasha Ravichander, Kyle Richardson, Zejiang Shen, Emma Strubell, Nishant Subramani, Oyvind Tafjord, Pete Walsh, Luke Zettlemoyer, Noah A. Smith, Hannaneh Hajishirzi, Iz Beltagy, Dirk Groeneveld, Jesse Dodge, Kyle Lo

Copyright and Artificial Intelligence: Policy Studies and Guidance 2024 misc

{U.S. Copyright Office}

Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model 2024 article

Ahmet Üstün, Viraat Aryabumi, Zheng-Xin Yong, Wei-Yin Ko, Daniel D'souza, Gbemileke Onilude, Neel Bhandari, Shivalika Singh, Hui-Lee Ooi, Amr Kayid, Freddie Vargus, Phil Blunsom, Shayne Longpre, Niklas Muennighoff, Marzieh Fadaee, Julia Kreutzer, Sara Hooker

Push and Pull: A Framework for Measuring Attentional Agency 2024 article

Wojtowicz, Zachary and Jain, Shrey and Vincent, Nicholas

A Systematic Review of NeurIPS Dataset Management Practices 2024 article

Yiwei Wu, Leah Ajmani, Shayne Longpre, Hanlin Li

Data-Sharing Markets: Model, Protocol, and Algorithms to Incentivize the Formation of Data-Sharing Consortia 2023 article

Raul Castro Fernandez

Alpaca: A Strong, Replicable Instruction-Following Model 2023 misc

{Stanford CRFM}

LEACE: Perfect linear concept erasure in closed form 2023 article

Nora Belrose, David Schneider-Joseph, Shauli Ravfogel, Ryan Cotterell, Edward Raff, Stella Biderman

Quantifying Memorization Across Neural Language Models 2023 inproceedings

Nicholas Carlini, Daphne Ippolito, Matthew Jagielski, Katherine Lee, Florian Tramer, Chiyuan Zhang

Understanding CC Licenses and Generative AI 2023 misc

{Creative Commons}

Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4 2023 article

Kent K. Chang, Mackenzie Cramer, Sandeep Soni, David Bamman

Wikipedia's value in the age of generative {AI} 2023 misc

Deckelmann, Selena

If there was a generative artificial intelligence system that could, on its own, write all the information contained in Wikipedia, would it be the same as Wikipedia today?

Algorithmic Collective Action in Machine Learning 2023 inproceedings

Moritz Hardt, Eric Mazumdar, Celestine Mendler-Dünner, Tijana Zrnic

Provides theoretical framework for algorithmic collective action, showing that small collectives can exert significant control over platform learning algorithms through coordinated data strategies.

ISO/IEC 23894:2023 Information Technology—Artificial Intelligence—Risk Management 2023 standard
A Watermark for Large Language Models 2023 article

John Kirchenbauer, Jonas Geiping, Yuxin Wen, Jonathan Katz, Ian Miers, Tom Goldstein

The Dimensions of Data Labor: A Road Map for Researchers, Activists, and Policymakers to Empower Data Producers 2023 article

Hanlin Li, Nicholas Vincent, Stevie Chancellor, Brent Hecht

Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment 2023 article

Yang Liu, Yuanshun Yao, Jean-Francois Ton, Xiaoying Zhang, Ruocheng Guo, Hao Cheng, Yegor Klochkov, Muhammad Faaiz Taufiq, Hang Li

Textbooks Are All You Need II: phi-1.5 technical report 2023 article

Yuanzhi Li, Sebastien Bubeck, Ronen Eldan, Allie Del Giorno, Suriya Gunasekar, Yin Tat Lee

SILO Language Models: Isolating Legal Risk In a Nonparametric Datastore 2023 article

Sewon Min, Suchin Gururangan, Eric Wallace, Weijia Shi, Hannaneh Hajishirzi, Noah A. Smith, Luke Zettlemoyer

Artificial Intelligence Risk Management Framework (AI RMF 1.0) 2023 techreport
OWASP Top 10 for Large Language Model Applications 2023 misc
TRAK: Attributing Model Behavior at Scale 2023 inproceedings

Sung Min Park, Kristian Georgiev, Andrew Ilyas, Guillaume Leclerc, Aleksander Madry

Introduces TRAK (Tracing with the Randomly-projected After Kernel), a data attribution method that is both effective and computationally tractable for large-scale models by leveraging random projections.

Terms-we-serve-with: Five dimensions for anticipating and repairing algorithmic harm 2023 article

Rakova, Bogdana, Shelby, Renee, Ma, Megan

Sociotechnical Harms of Algorithmic Systems: Scoping a Taxonomy for Harm Reduction 2023 inproceedings

Shelby, Renee, Rismani, Shalaleh, Henne, Kathryn, Moon, AJung, Rostamzadeh, Negar, Nicholas, Paul, Yilla-Akbari, N'Mah, Gallegos, Jess, Smart, Andrew, Garcia, Emilio, Virk, Gurleen

Understanding the landscape of potential harms from algorithmic systems enables practitioners to better anticipate consequences of the systems they build. It also supports the prospect of incorporating controls to help minimize harms that emerge from the interplay of technologies and social and cultural dynamics. A growing body of scholarship has identified a wide range of harms across different algorithmic technologies. However, computing research and practitioners lack a high level and synthesized overview of harms from algorithmic systems. Based on a scoping review of computing research (n=172), we present an applied taxonomy of sociotechnical harms to support a more systematic surfacing of potential harms in algorithmic systems. The final taxonomy builds on and refers to existing taxonomies, classifications, and terminologies. Five major themes related to sociotechnical harms — representational, allocative, quality-of-service, interpersonal harms, and social system/societal harms — and sub-themes are presented along with a description of these categories. We conclude with a discussion of challenges and opportunities for future research.

An Alternative to Regulation: The Case for Public AI 2023 article

Vincent, Nicholas, Bau, David, Schwettmann, Sarah, Tan, Joshua

The Dimensions of Data Labor: A Road Map for Researchers, Activists, and Policymakers to Empower Data Producers 2023 inproceedings

Li, Hanlin and Vincent, Nicholas and Chancellor, Stevie and Hecht, Brent

Power and Progress: Our Thousand-Year Struggle Over Technology and Prosperity| Winners of the 2024 Nobel Prize for Economics 2023 book

Simon Johnson, Daron Acemoglu

The Fallacy of AI Functionality 2022 article

Inioluwa Deborah Raji, Indra Elizabeth Kumar, Aaron Horowitz, Andrew D. Selbst

Common Crawl — Web-scale Data for Research 2022 misc

{Common Crawl}

Dataset Security for Machine Learning: Data Poisoning, Backdoor Attacks, and Defenses 2022 article

Micah Goldblum, Dimitris Tsipras, Chulin Xie, Xinyun Chen, Avi Schwarzschild, Dawn Song, Aleksander Madry, Bo Li, Tom Goldstein

Comprehensive survey systematically categorizing dataset vulnerabilities including poisoning and backdoor attacks, their threat models, and defense mechanisms.

Training Data Influence Analysis and Estimation: A Survey 2022 article

Zayd Hammoudeh, Daniel Lowd

DeepCore: A Comprehensive Library for Coreset Selection in Deep Learning 2022 article

Chengcheng Guo, Bo Zhao, Yanbing Bai

Comprehensive library and empirical study of coreset selection methods for deep learning, finding that random selection remains a strong baseline across many settings.

Datamodels: Predicting Predictions from Training Data 2022 inproceedings

Andrew Ilyas, Sung Min Park, Logan Engstrom, Guillaume Leclerc, Aleksander Madry

Proposes datamodels that predict model outputs as a function of training data subsets, providing a framework for understanding data attribution through retraining experiments.

Beta Shapley: a Unified and Noise-reduced Data Valuation Framework for Machine Learning 2022 inproceedings

Yongchan Kwon, James Zou

Generalizes Data Shapley using Beta weighting functions, providing noise-reduced data valuation that better handles outliers and mislabeled data detection.

LAION-5B: A New Era of Open Large-Scale Multi-Modal Datasets 2022 misc
Training language models to follow instructions with human feedback 2022 article

Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, Ryan Lowe

Probabilistic Machine Learning: An introduction 2022 book

Kevin P. Murphy

Releasing Re-LAION-5B 2022 misc
Why Black Box Machine Learning Should Be Avoided for High-Stakes Decisions, in Brief 2022 article

Cynthia Rudin

{LAION}-5B: An Open Large-Scale Dataset for Training Next {CLIP} Models 2022 inproceedings

Schuhmann, Christoph, Beaumont, Romain, Vencu, Richard, Gordon, Cade, Wightman, Ross, Cherti, Mehdi, Coombes, Theo, Katta, Aarush, Mullis, Clayton, Wortsman, Mitchell, Schramowski, Patrick, Kundurthy, Srivatsa, Crowson, Katherine, Schmidt, Ludwig, Kaczmarczyk, Robert, Jitsev, Jenia

Beyond neural scaling laws: beating power law scaling via data pruning 2022 article

Sorscher, Ben, Geirhos, Robert, Shekhar, Shashank, Ganguli, Surya, Morcos, Ari

The Stack: A Permissively Licensed Source Code Dataset 2022 misc

{BigCode Project}

Introducing Whisper 2022 misc

{OpenAI}

Robust Speech Recognition via Large-Scale Weak Supervision 2022 article

Alec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine McLeavey, Ilya Sutskever

On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 2021 inproceedings

Bender, Emily M., Gebru, Timnit, McMillan-Major, Angelina, Shmitchell, Shmargaret

Extracting Training Data from Large Language Models 2021 inproceedings

Carlini, Nicholas, Tramer, Florian, Wallace, Eric, Jagielski, Matthew, Herbert-Voss, Ariel, Lee, Katherine, Roberts, Adam, Brown, Tom B., Song, Dawn, Erlingsson, {\'U}lfar, Oprea, Alina, Papernot, Nicolas

Unsolved Problems in ML Safety 2021 article

Dan Hendrycks, Nicholas Carlini, John Schulman, Jacob Steinhardt

Beta Shapley: A Unified and Noise-Reduced Data Valuation Framework for Machine Learning 2021 article

Kwon, Yongchan and Zou, James

What's in the Box? An Analysis of Undesirable Content in the Common Crawl Corpus 2021 inproceedings

Alexandra Sasha Luccioni, Joseph D. Viviano

Measuring Mathematical Problem Solving With the MATH Dataset 2021 article

Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, Jacob Steinhardt

The Pile: An 800GB Dataset of Diverse Text for Language Modeling 2021 article

Leo Gao, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, Horace He, Anish Thite, Noa Nabeshima, Shawn Presser, Connor Leahy

What are you optimizing for? Aligning Recommender Systems with Human Values 2021 article

Jonathan Stray, Ivan Vendrov, Jeremy Nixon, Steven Adler, Dylan Hadfield-Menell

Quantifying the Invisible Labor in Crowd Work 2021 article

Carlos Toxtli, Siddharth Suri, Saiph Savage

Can "Conscious Data Contribution" Help Users to Exert "Data Leverage" Against Technology Companies? 2021 article

Nicholas Vincent, Brent Hecht

Data Leverage: A Framework for Empowering the Public in its Relationship with Technology Companies 2021 inproceedings

Vincent, Nicholas and Li, Hanlin and Tilly, Nicole and Chancellor, Stevie and Hecht, Brent

A Deeper Investigation of the Importance of Wikipedia Links to Search Engine Results 2021 article

Nicholas Vincent, Brent Hecht

Ethical and Social Risks of Harm from Language Models 2021 article

Weidinger, Laura, Mellor, John, Rauh, Maribeth, Griffin, Conor, Uesato, Jonathan, Huang, Po-Sen, Cheng, Myra, Glaese, Mia, Balle, Borja, Kasirzadeh, Atoosa, Kenton, Zac, Brown, Sasha, Hawkins, Will, Stepleton, Tom, Biles, Courtney, Birhane, Abeba, Haas, Julia, Rimell, Laura, Hendricks, Lisa Anne, Isaac, William, Legassick, Sean, Irving, Geoffrey, Gabriel, Iason

Language (Technology) is Power: A Critical Survey of “Bias” in NLP 2020 inproceedings

Blodgett, Su Lin, Barocas, Solon, Daum{\'e} III, Hal, Wallach, Hanna

Language Models are Few-Shot Learners 2020 article

Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei

Artificial Intelligence, Values, and Alignment 2020 article

Iason Gabriel

Lessons from Archives: Strategies for Collecting Sociocultural Data in Machine Learning 2020 inproceedings

Jo, Emily, Gebru, Timnit

Scaling Laws for Neural Language Models 2020 article

Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, Dario Amodei

Establishes power-law scaling relationships between language model performance and model size, dataset size, and compute, spanning seven orders of magnitude.

Exploring Research Interest in Stack Overflow -- A Systematic Mapping Study and Quality Evaluation 2020 article

Sarah Meldrum, Sherlock A. Licorish, Bastin Tony Roy Savarimuthu

Coresets for Data-efficient Training of Machine Learning Models 2020 inproceedings

Baharan Mirzasoleiman, Jeff Bilmes, Jure Leskovec

Introduces CRAIG (Coresets for Accelerating Incremental Gradient descent), selecting subsets that approximate full gradient for 2-3x training speedups while maintaining performance.

The Economics of Maps 2020 article

Abhishek Nagaraj, Scott Stern

Estimating Training Data Influence by Tracing Gradient Descent 2020 inproceedings

Garima Pruthi, Frederick Liu, Mukund Sundararajan, Satyen Kale

Introduces TracIn, which computes influence of training examples by tracing how test loss changes during training. Uses first-order gradient approximation and saved checkpoints for scalability.

The pushshift reddit dataset 2020 article

Baumgartner, Jason, Zannettou, Savvas, Keegan, Brian, Squire, Megan, Blackburn, Jeremy

Example Citation Placeholder 2020 misc

Jane Smith

Placeholder reference to support example citations in docs. Replace with a real source when available.

Are anonymity-seekers just like everybody else? An analysis of contributions to Wikipedia from Tor 2020 inproceedings

Tran, Chau, Champion, Kaylea, Forte, Andrea, Hill, Benjamin Mako, Greenstadt, Rachel

In Pursuit of Interpretable, Fair and Accurate Machine Learning for Criminal Recidivism Prediction 2020 article

Caroline Wang, Bin Han, Bhrij Patel, Cynthia Rudin

Reconciling modern machine-learning practice and the classical bias–variance trade-off 2019 article

Belkin, Mikhail, Hsu, Daniel, Ma, Siyuan, Mandal, Soumik

Common voice: A massively-multilingual speech corpus 2019 article

Ardila, Rosana, Branson, Megan, Davis, Kelly, Henretty, Michael, Kohler, Michael, Meyer, Josh, Morais, Reuben, Saunders, Lindsay, Tyers, Francis M, Weber, Gregor

The Secret Sharer: Measuring Unintended Memorization in Neural Networks 2019 inproceedings

Carlini, Nicholas, Liu, Chang, Erlingsson, {\'U}lfar, Kos, Jernej, Song, Dawn

Excavating AI: The Politics of Images in Machine Learning Training Sets 2019 misc

Crawford, Kate, Paglen, Trevor

Ecosystem Tipping Points in an Evolving World 2019 article

Vasilis Dakos, Blake Matthews, Andrew P. Hendry, Jonathan Levine, Nicolas Loeuille, Jon Norberg, Patrik Nosil, Marten Scheffer, Luc De Meester

Data Shapley: Equitable Valuation of Data for Machine Learning 2019 inproceedings

Amirata Ghorbani, James Zou

Proposes data Shapley as a metric to quantify the value of each training datum to predictor performance, satisfying equitable data valuation properties from cooperative game theory.

Data Shapley: Equitable Valuation of Data for Machine Learning 2019 inproceedings

Ghorbani, Amirata and Zou, James

Face Recognition Vendor Test (FRVT) Part 3: Demographic Effects 2019 techreport

Grother, Patrick, Ngan, Mei, Hanaoka, Kayee

Incomplete Contracting and AI Alignment 2019 inproceedings

Dylan Hadfield-Menell, Gillian K. Hadfield

HowTo100M: Learning a Text-Video Embedding by Watching Hundred Million Narrated Video Clips 2019 article

Antoine Miech, Dimitri Zhukov, Jean{-}Baptiste Alayrac, Makarand Tapaswi, Ivan Laptev, Josef Sivic

Towards Efficient Data Valuation Based on the Shapley Value 2019 inproceedings

Ruoxi Jia, Dah-Yuan Dao, Boxin Wang, Frances Ann Hubis, Nick Hynes, Neslihan M. Gurel, Carl J. Spanos

On the Accuracy of Influence Functions for Measuring Group Effects 2019 inproceedings

Pang Wei Koh, Kai-Siang Ang, Hubert H. K. Teo, Percy Liang

Privacy, anonymity, and perceived risk in open collaboration: A study of service providers 2019 inproceedings

McDonald, Nora, Hill, Benjamin Mako, Greenstadt, Rachel, Forte, Andrea

Model Cards for Model Reporting 2019 inproceedings

Mitchell, Margaret, Wu, Simone, Zaldivar, Andrew, Barnes, Parker, Vasserman, Lucy, Hutchinson, Ben, Spitzer, Elena, Raji, Inioluwa Deborah, Gebru, Timnit

Dissecting Racial Bias in an Algorithm Used to Manage the Health of Populations 2019 article

Obermeyer, Ziad, Powers, Brian, Vogeli, Christine, Mullainathan, Sendhil

Rosenbach v. Six Flags Entertainment Corp. 2019 legal
Fairness and Abstraction in Sociotechnical Systems 2019 inproceedings

Selbst, Andrew D., Boyd, Danah, Friedler, Sorelle A., Venkatasubramanian, Suresh, Vertesi, Janet

Mapping the Potential and Pitfalls of "Data Dividends" as a Means of Sharing the Profits of Artificial Intelligence 2019 article

Nicholas Vincent, Yichun Li, Renee Zha, Brent Hecht

"Data Strikes": Evaluating the Effectiveness of a New Form of Collective Action Against Technology Companies 2019 inproceedings

Nicholas Vincent, Brent Hecht, Shilad Sen

Simulates data strikes against recommender systems, showing that collective withholding of training data can create leverage for users against technology platforms.

Measuring the Importance of User-Generated Content to Search Engines 2019 inproceedings

Nicholas Vincent, Isaac Johnson, Patrick Sheehan, Brent Hecht

A Reductions Approach to Fair Classification 2018 inproceedings

Alekh Agarwal, Alina Beygelzimer, Miroslav Dudik, John Langford, Hanna Wallach

Should We Treat Data as Labor? Moving Beyond 'Free' 2018 article

Imanol Arrieta-Ibarra, Leonard Goff, Diego Jimenez-Hernandez, Jaron Lanier, E. Glen Weyl

Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science 2018 article

Bender, Emily M., Friedman, Batya

Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification 2018 inproceedings

Buolamwini, Joy, Gebru, Timnit

Datasheets for Datasets 2018 inproceedings

Gebru, Timnit, Morgenstern, Jamie, Vecchione, Briana, Vaughan, Jennifer Wortman, Wallach, Hanna, Daumé III, Hal, Crawford, Kate

The Dataset Nutrition Label: A Framework To Drive Higher Data Quality Standards 2018 misc

Sarah Holland, Ahmed Hosny, Sarah Newman, Joshua Joseph, Kasia Chmielinski

Troubling Trends in Machine Learning Scholarship 2018 article

Zachary C. Lipton, Jacob Steinhardt

A Blueprint for a Better Digital Society 2018 article

Weyl, E. Glen and Lanier, Jaron

Active Learning for Convolutional Neural Networks: A Core-Set Approach 2018 inproceedings

Ozan Sener, Silvio Savarese

Defines active learning as core-set selection, choosing points such that a model trained on the subset is competitive for remaining data. Provides theoretical bounds via k-Center problem.

Artificial Intelligence and Its Implications for Income Distribution and Unemployment 2018 techreport

Anton Korinek, Joseph E. Stiglitz

The Substantial Interdependence of Wikipedia and Google: A Case Study on the Relationship Between Peer Production Communities and Information Technologies 2017 inproceedings

McMahon, Connor and Johnson, Isaac and Hecht, Brent

Deep learning scaling is predictable, empirically 2017 article

Hestness, Joel, Narang, Sharan, Ardalani, Newsha, Diamos, Gregory, Jun, Heewoo, Kianinejad, Hassan, Patwary, Md, Ali, Mostofa, Yang, Yang, Zhou, Yanqi

The WARC Format 1.1 2017 misc

{International Internet Preservation Consortium}

Understanding Black-box Predictions via Influence Functions 2017 inproceedings

Pang Wei Koh, Percy Liang

Uses influence functions from robust statistics to trace model predictions back to training data, identifying training points most responsible for a given prediction.

Big Data's Disparate Impact 2016 article

Barocas, Solon, Selbst, Andrew D.

General Data Protection Regulation (EU) 2016/679 2016 misc

{European Union}

What's wrong with social simulations? 2014 article

Eckhart Arnold

The Algorithmic Foundations of Differential Privacy 2014 article

Cynthia Dwork, Aaron Roth

Children's Online Privacy Protection Rule (COPPA) — 16 CFR Part 312 2013 misc

{Federal Trade Commission}

The Future of Crowd Work 2013 inproceedings

Aniket Kittur, Jeffrey V. Nickerson, Michael Bernstein, Elizabeth Gerber, Aaron Shaw, John Zimmerman, Matt Lease, John Horton

Poisoning Attacks against Support Vector Machines 2012 inproceedings

Battista Biggio, Blaine Nelson, Pavel Laskov

Investigates poisoning attacks against SVMs where adversaries inject crafted training data to increase test error. Uses gradient ascent to construct malicious data points.

Guide to Protecting the Confidentiality of Personally Identifiable Information (PII) 2010 techreport

McCallister, Erika, Grance, Tim, Scarfone, Karen

Biometric Information Privacy Act (BIPA), 740 ILCS 14 2008 misc

{Illinois General Assembly}

Robust De-anonymization of Large Sparse Datasets 2008 inproceedings

Narayanan, Arvind, Shmatikov, Vitaly

Privacy as Contextual Integrity 2004 article

Nissenbaum, Helen

Modeling Complexity : The Limits to Prediction 2001 article

Michael Batty, Paul M. Torrens

HIPAA Privacy Rule — 45 CFR Parts 160 and 164 2000 misc

{U.S. Department of Health, Human Services}

Simple Demographics Often Identify People Uniquely 2000 article

Sweeney, Latanya

Social {Dilemmas}: {The} {Anatomy} of {Cooperation} 1998 article

Kollock, Peter

The study of social dilemmas is the study of the tension between individual and collective rationality. In a social dilemma, individually reasonable behavior leads to a situation in which everyone is worse off. The first part of this review is a discussion of categories of social dilemmas and how they are modeled. The key two-person social dilemmas (Prisoner’s Dilemma, Assurance, Chicken) and multiple-person social dilemmas (public goods dilemmas and commons dilemmas) are examined. The second part is an extended treatment of possible solutions for social dilemmas. These solutions are organized into three broad categories based on whether the solutions assume egoistic actors and whether the structure of the situation can be changed: Motivational solutions assume actors are not completely egoistic and so give some weight to the outcomes of their partners. Strategic solutions assume egoistic actors, and neither of these categories of solutions involve changing the fundamental structure of the situation. Solutions that do involve changing the rules of the game are considered in the section on structural solutions. I conclude the review with a discussion of current research and directions for future work.

No free lunch theorems for optimization 1997 article

Wolpert, David H, Macready, William G

The critical mass in collective action 1993 book

Marwell, Gerald, Oliver, Pamela

Family Educational Rights and Privacy Act (FERPA) 1974 misc

{U.S. Department of Education}

Alpaca Data Cleaned Repository misc

{gururise}

arXiv API User’s Manual misc

{arXiv.org}

arXiv OAI-PMH Interface misc

{arXiv.org}

arXiv Bulk Data Access misc

{arXiv.org}

Stanford Alpaca GitHub Repository misc

{Tatsu Lab}

C4 Generator Code misc

{TensorFlow Datasets}

Web Archiving File Formats Explained misc

{Common Crawl}

Common Crawl – Get Started misc

{Common Crawl}

Databricks Dolly Repository misc

{Databricks}

GSM8K Hugging Face Dataset Card misc

{OpenAI}

Grade-School Math (GSM8K) Repository misc

{OpenAI}

HH-RLHF Dataset misc

{Anthropic}

HowTo100M Project misc

{École Normale Supérieure}

Journal Article Tag Suite misc
JSON Lines Specification misc

{jsonlines.org}

WARC, Web ARChive file format misc

{Library of Congress}

Competition Math Dataset on Hugging Face misc

Hendrycks, Dan

Wikipedia Data Dumps – Dump Format misc

{Wikimedia Meta-Wiki}

NDJSON Specification misc

{ndjson}

OpenAssistant OASST1 Dataset Card misc

{OpenAssistant}

OpenAI API Reference – Chat misc

{OpenAI}

Apache Parquet Project misc

{Apache Software Foundation}

Project Gutenberg Offline Catalogs and Feeds misc

{Project Gutenberg}

Project Gutenberg File Formats misc

{Project Gutenberg}

Pushshift.io misc

{Pushshift}

Reddit API Documentation misc

{Reddit}

Reddit Data API Wiki misc

{Reddit Help}

Stack Exchange Data Explorer Help misc

{Stack Exchange}

Why is the Stack Exchange Data Dump only available in XML? misc

{Meta Stack Exchange}

BigCode Project Documentation misc

{BigCode Project}

The Stack dataset on Hugging Face misc

{BigCode Project}

The Stack v2 dataset on Hugging Face misc

{BigCode Project}

C4 dataset in TensorFlow Datasets misc

{TensorFlow Datasets}

TFRecord and tf.train.Example Tutorial misc

{TensorFlow}

Wikipedia Database Download misc

{Wikipedia}