How does computer science research move beyond academia and drive real-world innovation? This project presents a large-scale quantitative analysis of how computer science papers influence both patents and software development. By curating and cross-referencing a dataset of 200,000 papers from top conferences (AI, Computer Vision, NLP, and more), along with patent citations and GitHub repositories, we reveal two key pathways for research impact: (1) innovation, as measured by patents, and (2) development, via code repositories. We find that while only a minority of papers are cited by patents (9%) or used in repositories (11%), these papers receive a disproportionate share of citations, signaling outsized influence. Impact unfolds at different speeds—patent uptake can take 10–15 years, while code adoption happens rapidly, often within a year. Notably, research built on conventional knowledge is more likely to drive patents, whereas unconventional, novel combinations boost influence in software development. These insights provide a data-driven foundation for understanding and enhancing the real-world reach of computer science research.
Translational research, especially in the fast-evolving field of Artificial Intelligence (AI), is key to converting scientific findings into practical innovations. In Responsible AI (RAI) research, translational impact is often viewed through various pathways, including research papers, blogs, news articles, and the drafting of forthcoming AI legislation (e.g., the EU AI Act). However, the real-world impact of RAI research remains an underexplored area.
In this work, we aim capture RAI translational impact through two pathways: patents and code repositories, both of which provide a rich and structured source of data. Using a dataset of 200,000 papers from 1980 to 2022 in AI and related fields, including Computer Vision, Natural Language Processing, and Human-Computer Interaction, we developed a Sentence-Transformers Deep Learning framework to identify RAI papers. This framework calculates the semantic similarity between paper abstracts and a set of RAI keywords, which are derived from the NIST's AI Risk Management Framework; a framework that aims to enhance trustworthiness considerations in the design, development, use, and evaluation of AI products, services, and systems.
Key findings:
N.B.: If you do not receive the instruction message within a few hours, please check your junk/spam e-mail folder just in case the email was moved there.