Robohub.org
 

Funding high risk proposals at the National Science Foundation: A brief history


by
03 July 2013



share this:

National Science Foundation (NSF) efforts to develop a mechanism to fund research proposals that had a high risk of failure, but which also had the potential for high return, began in 1980. In that year a task force was created by the NSF Advisory Council to look at the issue of “highly creative or innovative” proposals for which there was “a high risk of failure.”

The task force’s report identified two significant hurdles the NSF needed to overcome to support high risk/high return proposals: 1) the conservatism of the peer review process, and 2) the reluctance of NSF program officers to fund research with a high potential for failure in an environment in which proposals with a high probability of success were not being funded because of limited budgets. The task force suggested a variety of ways to encourage NSF support of these risky proposals, most of which depended upon giving the program officer encouragement to use his or her discretion to identify and fund such proposals. None of the recommendations of the task force were immediately adopted by NSF, although some of the mechanisms suggested appear in modified form later.

A small number of failed grants were a necessary and unavoidable by-product of efforts to fund innovative research.

The first NSF mechanism for supporting high risk proposals was developed in the recently created Engineering Directorate in the mid-1980s. Engineering took the initiative because of the leadership of Assistant Director Nam Suh, who felt that a small number of failed grants were a necessary and unavoidable by-product of efforts to fund innovative research. The Directorate was also responding to recommendations by the National Academy of Engineering for NSF support of high-risk/high- return research. Engineering established a pilot project called Expedited Awards for Novel Research (EANR), encouraging program officers to fund innovative proposals without the benefit of peer review.

The successes of EANR funded proposals persuaded NSF Director Erich Bloch to establish the Small Grants for Exploratory Research (SGER) mechanism for the entire NSF in August 1989. Restricted to no more than 5% of any program’s budget, SGER awards, like EANR awards, bypassed the external peer review process. There were a number of objectives for SGER awards, but one was to induce researchers to “propose high-risk work to NSF.”

In 2004, the National Science Board established a Task Group on High-Risk Research to reexamine the issue. Its work, and that of an NSF staff task force on the topic, led to the replacement of the SGER award mechanism by a number of other mechanisms in 2009. The most important of these for high-risk research was the Early-concept Grant for Exploratory Research (EAGER). An EAGER award will provide up to $300,000 for up to two years of “high-risk, high-payoff” research. Like its predecessors, EAGER awards do not require external review.

References
National Science Foundation Advisory Council, “Report of Task Group 10: Funding of Innovative High Risk Proposals,” November 6, 1980
“NSF Starts Program to Encourage Creative Engineering Research,” NSF PR85-74, December 12, 1985
Erich Bloch, “Small Grants for Exploratory Research (SGER),” Staff Memorandum O/D 89-11, August 14, 1989
Dian O. Belanger, Enabling American Innovation: Engineering and the National Science Foundation, 1998, p. 246.
National Science Board, “Enhancing Support of Transformative Research at the National Science Foundation,” May 7, 2007
National Science Foundation, “Proposal Award Policies and Procedures Guide, Part I: Grant Proposal Guide,” October 2012



tags: , , , ,


Marc Rothenberg is agency Historian of the National Science Foundation.
Marc Rothenberg is agency Historian of the National Science Foundation.





Related posts :



Robot Talk Episode 119 – Robotics for small manufacturers, with Will Kinghorn

  02 May 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Will Kinghorn from Made Smarter about how to increase adoption of new tech by small manufacturers.

Multi-agent path finding in continuous environments

  01 May 2025
How can a group of agents minimise their journey length whilst avoiding collisions?

Interview with Yuki Mitsufuji: Improving AI image generation

  29 Apr 2025
Find out about two pieces of research tackling different aspects of image generation.

Robot Talk Episode 118 – Soft robotics and electronic skin, with Miranda Lowther

  25 Apr 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Miranda Lowther from the University of Bristol about soft, sensitive electronic skin for prosthetic limbs.

Interview with Amina Mević: Machine learning applied to semiconductor manufacturing

  17 Apr 2025
Find out how Amina is using machine learning to develop an explainable multi-output virtual metrology system.

Robot Talk Episode 117 – Robots in orbit, with Jeremy Hadall

  11 Apr 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Jeremy Hadall from the Satellite Applications Catapult about robotic systems for in-orbit servicing, assembly, and manufacturing.

Robot Talk Episode 116 – Evolved behaviour for robot teams, with Tanja Kaiser

  04 Apr 2025
In the latest episode of the Robot Talk podcast, Claire chatted to Tanja Katharina Kaiser from the University of Technology Nuremberg about how applying evolutionary principles can help robot teams make better decisions.

AI can be a powerful tool for scientists. But it can also fuel research misconduct

  31 Mar 2025
While AI is allowing scientists to make technological breakthroughs, there’s also a darker side to the use of AI in science: scientific misconduct is on the rise.



 

Robohub is supported by:




Would you like to learn how to tell impactful stories about your robot or AI system?


scicomm
training the next generation of science communicators in robotics & AI


©2025.05 - Association for the Understanding of Artificial Intelligence


 












©2025.05 - Association for the Understanding of Artificial Intelligence