Request for Proposal - Bill & Melinda Gates Foundation
Transcription
Request for Proposal - Bill & Melinda Gates Foundation
New Approaches for Evaluating and Improving the Efficacy of Digital Courseware: Request for Proposal Application Deadline: March 20, 2015 – 5:00 pm (PST) Overview Introduction In 2013, The Bill & Melinda Gates Foundation invested $2 million in new approaches to facilitating short-cycle feedback so that teachers, school decision-makers, and parents have better information about the effectiveness of digital courseware, and developers of such products could receive rapid input to improve their offerings. That investment supported the launch two types of new initiatives: Test bed networks, organizations that support schools in piloting innovative courseware products and help schools and entrepreneurs to collaboratively define a structured process to evaluate and improve the efficacy of digital courseware; and Efficacy platforms, technology platforms that use validated diagnostic and formative assessments to facilitate rapid iteration and improvement of digital content and tools. To date, these initiatives have supported more than 50,000 students and 2,000 teachers in piloting new courseware. Based on that success, we are requesting proposals for a second cohort of test bed networks and efficacy platforms that can help improve digital courseware’s efficacy in: 1. Student Learning: effective digital courseware should not only enable students to achieve explicit learning objectives but also accelerate the pace at which they do so. 2. Teacher Support: effective digital courseware should be easy to implement and integrate into instructional plans, saving teachers time and helping them support their students’ learning needs. 3. Engagement: effective digital courseware should provide for a rich, compelling, and interactive experience that will encourage repeated, prolonged, and self-motivated use. 4. Satisfaction: effective digital courseware should sufficiently meet student and teacher needs and provide for a consistent, inviting, and intuitive experience. This document is an invitation for proposals for rapid digital courseware evaluation and improvement approaches that: Focus the development and adoption of personalized learning products on helping students achieve desired learning outcomes. Put teachers and school decision-makers at the center of the shift toward personalized learning. Lower risks and barriers to all parties of adopting new, potentially transformative products. Encourage the rapid development of a healthy, transparent market for highly effective learning technologies. Page 1 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Vision for Personalized Learning Our vision for education is that all students have learning experiences that are tailored to their individual needs, skill levels, and interests. For some, this will mean accelerating instruction beyond the fixed pace of many of today’s classrooms. For others it will mean being given opportunities for extended practice of skills or content they have yet to master. For all students, it will mean learning experiences that are tailored specifically to their progress against a clear set of academic standards; receiving constant, actionable information on their progress; experiencing more one-on-one engagement with their teacher and developing deep connections with other students. Implementing this vision, which we and many others call personalized learning, at scale will require new tools to help teachers, parents, and students diagnose gaps in knowledge and skills and adapt learning experiences based on student progress. Technology can help support effective teachers, by giving them new ways to design and provide personalized instruction to their students. Comprehensive courseware products such as Carnegie Learning’s Cognitive Tutor, Reasoning Mind, ST Math, and Dreambox are facilitating significant learning gains for students in even the highest-need schools, while apps like Refractions, DragonBox, and Motion Math are demonstrating that even rigorous academic content can be highly engaging. Our goal is for all students to have ready access to these types of tools, enabling them to learn what they need, when they need it by personalizing content, learning experiences, and support in real time. Market Context Fueled by the growth of personalized learning schools, technological advances, and an influx of new talent, “start-up companies and individuals are developing digital learning resources at a rapid pace.” (US Department of Education, 2013). At the same time, tech-savvy teachers are increasingly seeking products to support their efforts to personalize learning and improve student outcomes. More than 90% of all teachers in a recent Harris Interactive survey indicated that “they would like to use more education technology in the classroom than they do now” (Harris Interactive, 2013); however, the market is not yet rich enough with choices for pioneers implementing personalized learning, nor is information on the impact of digital courseware on student learning readily available enough to meaningfully drive purchasing decisions or to support the transition to a performance-driven market. As a result, many incumbent providers are able to compete based on the strength of their sales and marketing efforts alone, while innovative players with promising products struggle to achieve sustainability and scale. We envision a healthy, rational US K-12 digital courseware market characterized by effectiveness, affordability, and transparency. In such a market, vendors would have ready access to short-cycle efficacy data and information that would be used to drive continuous product improvement. In turn, teachers and school decision-makers would use standardized, easily comparable efficacy data in deciding whether to continue using existing products or to adopt new ones. Making this data and information readily available would help evolve the market to one in which school decisionmakers are well supported in identifying the most effective tools for student learning, vendors develop and compete primarily based on performance, and all students have ready access to highly effective personalized learning tools. One way in which we support these types of market shifts is by providing resources that cast a light on gaps and opportunities that are critical to realizing the vision of personalized learning for all students. Over the past year, our analysis of the market – informed by input from leading teachers and school decision-makers, innovative vendors, and a range of experts – identified three major shifts required to move towards a transparent, performance-based, and responsive K-12 digital courseware market: 1. A dramatic increase in the number of effective personalized learning products, particularly in high-need areas such as literacy. While investment in education has increased dramatically over the past several years, critical areas of the market remain unserved. For example, of the $427 million in venture funding provided for K-12 education entrepreneurs in 2012, only 6% of it was invested in digital instructional content (NewSchools Venture Fund, 2012). 2. Providing readily available, independent information on learning technologies that allow teachers and school decision-makers to discover new products, quickly identify product alignment to standards, and share feedback on learning technologies based on personal experience. Nearly 80% of teachers agree that “it’s at least somewhat difficult to find out about high quality educational technologies,” and nearly a third of all teachers spend an hour or more each week researching which tools to use (Harris Interactive, 2013). 3. The emergence of lightweight, rapid feedback loops between teachers, school decision-makers and entrepreneurs that use student learning and engagement data to drive decision-making and continuous improvement. The federal Page 2 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal What Works Clearinghouse was designed to show what programs and products are most effective. However, due to its methodological requirements, it does not sufficiently address the needs of schools or vendors. Increasing the speed and transparency with which valid data is produced and made available should increase incentives for vendors to improve the performance of their products. Our Literacy Courseware Challenge invested $6 million to help facilitate the first shift. Our support of organizations that provide ready information on digital learning products, such as Graphite, EdReports, and EdSurge, seeks to facilitate the second shift. This proposal will seek to facilitate the third shift. Thus, we are requesting proposals for new approaches to facilitating short-cycle feedback on product efficacy to both developers of innovative, digital courseware and the students, teachers, and school decision-makers they serve. Detailed information on the problem we seek to solve, our definition of product efficacy, grant awards, and eligibility is contained within this request for proposal. Scope and Approach This request for proposal will support a three-year grant program targeting product entrepreneurs, market catalysts, and school networks, or any partnerships or combinations of these parties. The Bill & Melinda Gates Foundation does not envision being a sole funder of these initiatives during the two-phase process or thereafter. Applicants should have the ability and inclination to generate additional funds and to support on-going operations. Solution Requirements This RFP was developed by the Bill & Melinda Gates Foundation with the input and guidance of an advisory group composed of eight innovative school and non-profit leaders, and entrepreneurs. While the content and viewpoints expressed herein are those of the foundation, the vision and solution requirements for the RFP were shaped through extensive research on existing solutions and in-depth interviews with the advisory group. Throughout 12 interviews, advisory group members consistently described product development and piloting as an ambiguous, time-consuming, and often isolated process that poses large risks to schools and creates a very high barrier to entry for entrepreneurs with potential solutions. Faced with an increasing number of product options, school decisionmakers reported both a lack of data on what products work for which students under which circumstances and a lack of means to quickly evaluate whether a given tool is the right fit for them – a particular concern at a time when lean school budgets have only heightened the need to justify and assess the return on any investment in technology. Entrepreneurs, in turn, cited barriers including limited feedback channels and an insufficient number of teachers and students to building effective products that satisfy school needs. Both groups agreed that currently available data sources were insufficient for evaluating product efficacy, in particular a product’s impact on student learning. Among the major limitations they cited are: The rigorous effectiveness studies that have traditionally produced this data are time-consuming, expensive, and unable to keep pace with the abundance of emerging tools. For example, the federal What Works Clearinghouse lists fewer than 50 K-12 literacy interventions that show positive or potentially positive impact on student learning. Data produced by existing methods is often based on infrequent interim or end-of-year tests that assess student performance at a single point in time, forcing school decision-makers to rely on data that is inconsistently available and lags far behind ongoing product improvements. Product efficacy data from credible sources typically provides only a single dimension of product quality and statements about large, general populations. This type of data is useful in assessing a product independent of context, but is not helpful to school decision-makers who need to understand how a particular product is affecting an individual student’s learning and take immediate action based on this information. Self-reported vendor claims of efficacy cannot be easily evaluated and corroborated. This leads to even validly produced studies being viewed with distrust, and schools in some cases unnecessarily duplicating efforts by contracting with independent evaluation firms to conduct their own efficacy studies. Page 3 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Vendor-produced efficacy data is often presented in unwieldy, hard-to-interpret, and inflexible formats, eroding its usefulness for already-oversubscribed school staffs. School Needs In a recent survey of more than 1,000 teachers, the most cited reason for frustration with technology in the classroom was that “they don’t do exactly what I want” (EdShelf, 2013). Still, nearly all teachers “agree they would like to use more education technology than they currently do,” citing its potential positive impact on student outcomes and personalizing learning (Common Sense Media, 2013). In order to achieve this, schools need technologies that are both efficient and effective at delivering personalized learning experiences, and that make the best use of school and teacher time, effort, and expertise. More than ever, teachers and school decision-makers need practical, credible evidence on the effectiveness of new tools – and that evidence’s applicability to their particular needs before adopting the tools. As a part of that evidence, they need to understand how to implement new tools based on prior success, their unique student needs, and fit between test conditions and real classroom environments. Post-implementation, teachers need regular data showing whether the tools they are using are having a positive impact on student learning and engagement. If something is not working optimally, they need to have at their disposal built-in feedback channels with entrepreneurs to help diagnose and address those deficiencies through product refinements in a timely manner. Vendor Needs Entrepreneurs tell us that these shortcomings pose a Catch-22: their products are often not trusted by schools due to a lack of independent, long-term evidence of efficacy – but they also cannot get their products used by a critical mass of people to generate the evidence that they need. Furthermore, while entrepreneurs are increasingly adopting an agile product development process that creates opportunities to refine and strengthen products in response to feedback in a matter of weeks, the lack of real-time, product efficacy data makes this process at best one guided primarily on engagement alone rather than actual student learning. As a result, some entrepreneurs build products that are engaging enough for individual teachers to employ in classrooms as an instructional change of pace, but that are not optimized for student learning. When successful, this approach generates an initial surge in use, but often fails to build an evidence base that justifies wider deployment. Other entrepreneurs focus immediately on demonstrating impact on student learning, but struggle either to test their products with enough students to generate statistically significant data or to show clear, positive impact without incurring major expenses from hiring independent evaluators. In both cases, innovative entrepreneurs struggle to establish themselves in the market and to compete against incumbents. Entrepreneurs, therefore, need a critical mass of pilot teachers and students to help inform and validate their product efforts. They also need to be able to lower the risk of trying new, untested products by quickly proving to schools that they have the processes in place to gather and provide credible, actionable efficacy data at intervals short enough to support short-cycle adjustments to student learning experiences without compromising the entrepreneur’s own ability to test and refine products. As adoption increases, entrepreneurs need robust mechanisms within their product development processes to continuously measure efficacy, to learn about the effects of different implementations on student learning and engagement, and to rapidly iterate and optimize their products in response to both qualitative and quantitative feedback. Rapid cycles of modification, analysis of results, and redesign are “key to producing dramatic change while reducing risk” (Tony Bryk, 2011). Solutions Requested To address these needs, we are requesting proposals for test bed networks, platforms that utilize A/B testing with validated assessments, or solutions combining the two. Successful proposals will address at least one of the solution descriptions below and will rate highly against the evaluation criteria below. Test Bed Networks Forums that bring teachers and entrepreneurs together in the hopes of facilitating collaborative product development are increasingly being convened across the country by organizations such as the Education Technology Start-up Collaborative. These events typically feature teachers “test-driving” personalized learning products across various stages of development for free or at a reduced price and shaping the product development process by offering valuable feedback to entrepreneurs. Events such as these are a step in the right direction, but they are limited in continuity and scope of impact. A sustained test bed network with consistent and frequent interactions between teachers, their Page 4 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal students, and entrepreneurs on a larger scale – in either an online or offline medium – can facilitate a meaningful feedback loop about the qualitative and quantitative dimensions of product efficacy. For entrepreneurs, these networks would provide a range of benefits that would help them compete more effectively in the market and increase their chances of success, including: access to a critical mass of pilot teachers and students; the feedback necessary to build more effective products; the ability to generate early evidence of their products’ impact on student learning and engagement, and greater teacher satisfaction. For schools, these networks would create a way to quickly determine which products merit further exploration, greater influence on product design, and better, more regular data about the effectiveness of the learning technologies used in their classrooms. As a part of these test bed networks, schools and entrepreneurs would collaboratively define a structured process to evaluate and improve the efficacy of digital courseware, including: (1) the implementation model for each product tested, (2) the format and frequency of any requisite professional development, (3) the specific measures of product efficacy to be evaluated and by whom, (4) the diagnostic and assessment to be administered, (5) the nature of data reporting and data-sharing agreements between school and vendor, and (6) the format and structure of the feedback loop. Platforms (Utilizing A/B Testing with Validated Assessments) While software programs such as Google Analytics and Mixpanel have provided an effective means by which to digitally measure engagement and satisfaction in web-based products, the largest gap when it comes to evaluating product efficacy remains in measuring a product’s impact on student learning. The randomized controlled trial (RCT) has long been recognized in many fields, including education, as the gold standard for evaluating the efficacy of an intervention. While this experimental model is viewed by researchers as the most conclusive means by which to establish a causal relationship between a given practice and a learning outcome, in its traditional form it poses a number of often-insurmountable challenges in schools, including in a digital learning context. Namely, RCTs can be cost-prohibitive for resource-starved schools and entrepreneurs, costing hundreds of thousands of dollars (Coalition for Evidence-Based Policy, 2012), and can often take years to fully execute, a timeline which is incompatible with the velocity of change and decision-making in technology and education. Random assignment of students to treatment and control groups is also very difficult to accomplish, given the way schools typically assign students to teachers and programs. For these reasons, fully leveraging the transformative potential of innovative, personalized learning products must first involve “accepting that the strongest level of causal evidence is not necessary for every learning resource decision" (US Department of Education, 2013). A/B testing, a widely-practiced product development heuristic employed by many education entrepreneurs, captures the essence of random assignment in RCTs and allows for systematic comparison of different product features – without the prohibitive cost or timeline required. Under this method, two randomly-assigned groups of students are given versions of the product that vary in a single, deliberate manner – thus, allowing comparisons to be drawn between the experiences and outcomes of the two groups, and influential product changes identified. For early-stage entrepreneurs with new products and teacher and student bases insufficient for large-scale, multi-site RCTs and schools with short decision-making timelines and low tolerance for risk, a collaborative form of A/B testing utilizing validated diagnostic and formative assessments represents the ideal vehicle with which to carry out small-scale efficacy studies. The information gleaned from these studies can serve as crucial inputs into the product development and purchasing processes in the near-term by proving that impacts on student learning are maintained even outside of the specific product environment. In the long-term, it will also form the foundation upon which larger-scale experiments can be performed to demonstrate fidelity of results across a broader diversity of teachers, students, and contexts. Page 5 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Exhibit 1: Evolution of Efficacy Studies Use Cases The scenarios depicted below are representative of our overall vision (Phase I + Phase II). They are presented here as potential examples of proposals to help guide potential respondents only and should not be perceived as prescriptive or limiting in any fashion. Test Bed Networks Better Schools for All (BSA), a non-profit organization that supports the expansion of high-performing charter schools, submitted an application for Phase I of the New Approaches to Evaluating and Improving the Efficacy of Digital Courseware RFP. BSA, whose network of twenty schools nationally serves a population of over 8,000 students, had been thinking about a systematic means of evaluating the efficacy of a seemingly ever-increasing pipeline of new digital learning products up for consideration for use in classrooms. Intrigued by the vision of a test bed network that would facilitate the structured piloting of digital courseware in classrooms, connect teachers and students with entrepreneurs in a mutually-beneficial way, and ultimately enable rapid decision-making based on product efficacy data, BSA formed a working team composed of a select group of entrepreneurs, school decision-makers, and teachers, to begin exploring a potential solution. Over a series of planning meetings, the team agreed upon the following parameters of the test bed network: Implementation format for each product (station rotation blended model, 30 minutes a session, two times per week) Professional development format and frequency (half-day, in-person, vendor-facilitated overview combined with justin-time online PD as needed) Efficacy metrics to be tracked and evaluated (impact on student learning, teacher support, student engagement, and student/teacher satisfaction) The external measure of impact on student learning (a standards-aligned, initial diagnostic and formative assessment already in use presently) The content, level of detail, and visual format of data outputs for all stakeholders The structure of the feedback loop (weekly reports on student engagement, monthly reports on student/teacher satisfaction as measured by surveys, monthly reports on impact to student learning, real-time IT/trouble-shooting support) The team applied for and won a Phase 1 grant for a six-month pilot of the solution for three new literacy products, across ten sixth-grade classrooms within the BSA network. Page 6 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal For entrepreneurs, the pilot generated a continuous stream of valuable feedback from real teachers and students that could be used to improve the product experience. It also exposed inconsistencies between how the products were designed to be implemented and how they were actually being utilized in the classroom. For teachers, the pilot provided a way to quickly assess how effective new products actually are in the classroom, and an opportunity to help shape the product development process as well. For school decision-makers, the pilot revealed significant and persistent relative differences in student learning gains and engagement across the three products, allowing them to make faster decisions about which products to explore further or discontinue. Based on the success of this initial pilot, BSA is currently seeking Phase II funding to deploy the solution across its entire portfolio of schools. Platforms In hopes of building a decision-making tool for teachers and school decision-makers to quickly evaluate the efficacy of digital courseware products in a systematic fashion, DecisionEngine, an enterprise software company specializing in data analytics, submitted an application for Phase 1 of the New Approaches to Evaluating and Improving the Efficacy of Digital Courseware RFP. The proposal was for ProductAssess, a web-based evaluation system that would: Integrate with web-based products so as to be able to measure student learning and engagement resulting from use of those products Create a student portal within which students would be able to securely log-in, access, and use the digital products under evaluation, and subsequently complete standards-aligned diagnostics and formative assessments Create teacher/administrator/parent portals within which stakeholders would be able to access actionable data on student engagement and learning gains in short cycles, and in both granular and aggregate levels of detail as needed Provide continuous feedback on product efficacy and fidelity of implementation to personalized learning vendors, who can then quickly translate the feedback into product refinements/enhancements Have the capacity to perform automated randomized assignment tests on a large scale After winning the Phase I grant and building out a beta version of ProductAssess, DecisionEngine was able to secure a six-month pilot with a large, urban public school district and a subset of its portfolio of product vendors. Though this initial pilot was insufficient in scope to generate any statistically-definitive conclusions, the data suggested that there were indeed significant relative differences in student engagement and learning gains across the various digital learning products currently employed by the district. It also established for school decision-makers a standardized framework with which to measure product efficacy, provided useful feedback for ProductAssess improvements back to DecisionEngine, and provided information regarding fidelity of implementation back to product vendors. DecisionEngine is currently seeking Phase II funding for a national launch. With the efficacy data gleaned from a broader cross-section of schools and vendors, DecisionEngine hopes to utilize the full capability of ProductAssess to conduct automated randomized assignment tests in order to: Quickly assess the efficacy of newly-introduced personalized learning tools A/B test new product features/enhancements Recommend different implementations of products based on data from other schools Page 7 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Rules & Guidelines Eligibility Funding Criteria For this RFP (Phase I), we are targeting product entrepreneurs, market catalysts, and school networks/districts - or partnerships or combinations of these parties – with proposals for the following: Test Bed Networks: collaborative partnerships between schools and entrepreneurs that facilitate broad usage of digital courseware by teachers and their students in order to build a continuous feedback loop on product efficacy for school decision-makers and entrepreneurs. Platforms: technology products that will support data collection on digital courseware in order to assess product efficacy and that can support large-scale, short-cycle A/B testing. Combination: solutions that combine Test Bed Networks with Platforms. While we anticipate that most respondents will seek to address either a platform or test bed network in their proposals, special consideration will be given to organizations that – whether on their own or through proposed or already established partnerships – can deliver on a comprehensive solution that combines both. We will consider only those proposals that adequately address all four components of product efficacy as defined in this document, with the heaviest emphasis on impact on student learning. We encourage for- and non-profit organizations with demonstrated capacity for or promise of undertaking the work described herein, to respond. Teams proposing partnerships are required to submit a single application and identify the financial recipient of the grant funds. Given the nascent role of product efficacy efforts in the K-12 education ecosystem currently, we will evaluate and grant awards to ideas both in the Concept Development and Testing/Deployment stages. Type of Award Stage of Development Concept Development New concepts pre-beta testing Definition In use by <10 teachers and 300 students Size of Award Number of Awards Grant Period Up to $100K Up to 6 12 months Purpose of Grant To facilitate concept development and testing in preparation for pilot Testing/Deployment School-ready products In use by >10 teachers and 300 students and poised to grow Up to $350K Up to 4 12 months To facilitate solution refinement and pilot in preparation for broad implementation and independent, third-party evaluation Do I retain intellectual property ownership? Yes, all entrants, regardless of winner status, will retain intellectual property ownership of their software applications. However, winners must agree to support global access objectives by conducting and managing the project and information in a manner that enables (a) the knowledge gained during the project to be promptly and broadly disseminated and (b) the intended products to be made available and accessible for free or at reasonable cost to the people most in need. For more details, please see the Representative Terms and Conditions that will be included in the grant agreement. Page 8 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Exclusion Criteria We will NOT consider funding for: Respondents that do not propose either a platform and/or test bed network solution Proposals that fail to adequately address all four dimensions of product efficacy (as defined in this document) Single LLCs, or individuals not affiliated with a for- or non-profit company o If you are an LLC, please provide evidence that you're taxed as a partnership, or that you've made an election to be taxed as a corporation Respondents without an existing, minimally-viable solution Individual schools Isolated, one-off events bringing together entrepreneurs and school leaders (e.g. “School Start-Up Days”) Proposals with budgets >$350,000 or >12 months in duration Evaluation Criteria Submissions received by the deadline will be evaluated on the following criteria and rubric: Scoring Criteria Alignment with Objectives Potential for Impact Capacity to Deliver TOTAL Points 25 50 25 100 Alignment with Objectives (25%) o Fully addresses the problem statement and vision raised herein o Proposes to adequately and comprehensively measure all dimensions of product efficacy (as per the definition described herein) o Will serve a diverse cross-section of schools, teachers, and students o Is affordable and accessible to all schools Potential for Impact (50%) o Establishes robust, continuous, and scalable feedback channels and processes between stakeholders o Integrates diagnostics and formative assessments that are broadly-applicable and have been validated by instructional leaders and the research community o Includes short-cycle reporting interfaces and outputs that are user-friendly, flexible, intuitive, and actionable o Proposes a realistic implementation plan that requires minimal training and IT support by vendors and/or pilot schools o Utilizes technology that is interoperable, secure, scalable, and has the analytics capability to perform largescale A/B testing (if applicable) Capacity to Deliver (25%) o Led by a team with the skills and experience necessary for the proposed work o Proposes a work plan that can be realistically executed on time and within budget o Attests to the long-term financial health, sustainability, and self-sufficiency of the organization o Engenders confidence that the team has the human and financial capital to viably scale the solution if it succeeds in pilot testing o Demonstrates buy-in from proposed partners (if applicable) Page 9 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Activities & Timeline February 18, 2015: Request for Proposal (RFP) Open. February 24, 2015 – 1:00 PM (PST): Q&A call/webinar. March 20, 2015 – 5:00 PM (PST): Completed applications due. April 2015: Finalists notified. June 2015: Countersigned Grant Agreements received. Note: Grant awards will be contingent on the successful completion of due diligence processes. How to Apply Response Requirements To apply, please submit a proposal using the ‘Proposal Narrative’ template that includes the following: Summary (1/4 page or less): o Award category: Concept Development or Testing/Deployment o Amount requested ($) o Brief description: 1-2 sentence summary of your proposal Organization (3 pages or less): Please tell us about the people and organization(s) that will develop the solution, and provide: o Background about how long the organization has been in operation, its mission, and any products that are already available o A list and description of key existing or pipeline users of the solution o Historical information on number of existing users currently supported and anticipated growth trajectory o A list of all relevant team members complete with biographical information. Describe how your team members know one another and specify the activities each person will be involved in o An explanation for why this team has the unique capability to develop and implement this solution Solution Description (5 pages or less): Please provide a detailed description of the platform and/or process you aim to develop, and address the following: o What is a use case that accurately depicts how you envision your solution to work? o What short- and long-term goals do you propose for measuring and evaluating the success of your solution? o What specific metrics of product efficacy do you seek to measure, facilitate a feedback loop around, and optimize for? o What technologies will the solution utilize, and how will such technologies support facilitating a short-cycle feedback loop around product efficacy? o What assessments and/or feedback channels will be employed in the solution? To what extent have the assessments been validated by instructional leaders and the research community? What quantitative and qualitative feedback will teachers, school decision-makers, and entrepreneurs receive based on these assessments and channels, how, in what form, and how often? o What stakeholders and/or partners need to be involved and what is the desired nature, number, frequency, and structure of intercepts linking them together? o How do you specifically plan to deliver feedback on product efficacy to drive continuous improvement efforts for vendors? o If a platform, service or product, how will you price it? If a process, what rubric will you employ to determine who gets to participate and when? o What training, professional development, and ongoing support will be required to enable a pilot? o What is your cost to develop the platform? o How does your pricing compare to similar existing products? Additionally, respondents may provide up to five additional pages of appendices that specifically relate to solution description, such as screenshots, sample assessments, etc. Page 10 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Work Plan (1 page or less): Please outline the major activities associated with the development of the proposed solution, when they will be completed, and by what resources. Recruitment Plan (1 page or less): Please tell us how you plan to attract customers of the solution to demonstrate proof of concept. Include specific information on partnerships, distribution channels and/or marketing strategies. Submission Instructions Start by clicking the application link below. The first time you visit the portal you will need to create an account. After you enter your information, a temporary password will be emailed to you. Return to the portal to personalize your password and complete your account profile. 1. Click on the Application Portal link. 2. Create an Account 3. Once you are logged in, select “If you have received an invitation… you may respond here.” on the home page and enter the following information: Request ID: SOL1128135 Access Code: educ@tion2015 Please Note: You can save the application and return to it later by logging in to the Application Portal at any time before the submission deadline with your email address and password. The Request ID and Access Code are only necessary the first time you access the application. You can only hit “submit” once using the application tool, so we suggest iterating on documents offline and submitting your final versions. You will receive a confirmation email once you have successfully submitted your full application materials. If you do not receive the email (check your spam folder), please return to the portal and try again. To view frequently asked questions about submitting your application, click Help on the Application Portal. We use the application portal mainly to create an entry in our internal database. After you use it once to submit your documents, you’ll send all subsequent documents as attachments via email. The Application Portal will be updated periodically with information regarding the Q&A Call/ Webinar currently scheduled for February 24, 2015 – 1:00 PM (PST) Filling out the Online Application Follow the directions on the application portal. 1. Getting Started tab - Please review the all Guideline and Reference documents noted below. Guidelines and Reference Document New Approaches for Evaluating and A Improving the Efficacy of Digital Courseware - RFP B BMGF Representative Terms and Conditions C BMGF College Ready Strategy D BMGF Indirect Cost Policy E ER Guidelines F Data Stewardship Principles Page 11 of 15 Purpose Invitation to participate and guidelines for the RFP Provides the Representative Terms and Conditions for the grant. If you’re proposal is selected, you should be prepared to agree to the terms as stated in this document. Please review it carefully. Provides information about the Gates foundation USP College Ready Strategy Provides guidance about the foundation’s Indirect Cost policy Provides guidance on private foundation expenditure responsibility requirements Provides guidance on private foundation stewardship principles to protect student data and privacy © 2013 Bill & Melinda Gates Foundation Request for Proposal 2. Describe Project tab - Please fill this out to the best of your ability, using the following guidelines: Project Title – a short title, including the name of the solution. Project Description – a short description about the solution and any unique features. Project Duration – 12 months. Project Amount – enter the amount you are applying for. Please consult the RFP Information above for guidance on the maximum amount for each award. Referred By – please note where you heard about this RFP. 3. Upload Files tab - Please complete and upload the documents noted below using the templates provided. Templates and Organizational Documents Document Purpose Organization’s Current Year Budget Narrative Guidelines and template for narrative responses Captures broad budget and spending categories for the work Identify how the proposed work under the Project will fulfill the foundation’s Global Access objectives A collection of information about your organization (Note: Please gather the documents below into PDF form, and then submit as one document) A copy of your IRS tax determination letter or certification of formation (depending on what type of entity you are), as different grant terms and conditions may apply Applicable only to LLCs. If you are an LLC, please provide evidence that you are taxed as a partnership, or that you have made an election to be taxed as a corporation Audited financial statements for the past two years, or a proxy if you have not had an audit Your organization’s current-year budget List of Board of Directors List of your board of directors, if applicable 1 Proposal Narrative 2 Budget 3 IP Due Diligence Module 4 Organizational Documents IRS Tax Determination Letter or Certificate of Formation Evidence of Partnership Last 2 Years of Audited Financial Statements (or proxy) 4. Add Contacts tab - Please make sure you add a contact for the person who will sign the grant agreement if you are awarded a grant in order to facilitate the award process. Help Contact(s) Please contact college.ready@gatesfoundation.org for questions. Selection and Award Process Initial Application Review After all applications are submitted, Foundation staff will screen submissions to address whether the proposal addresses the key needs described in the topic and adheres to the application requirements. Applications excluded during the screening process will be notified that their proposals were declined. Page 12 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Application Review, Scoring, and Finalist Notification Next, a panel made up of Foundation and external reviewers will review submissions and issue a preliminary scoring based on the scoring rubric. At this point, a select number of organizations will be notified that they have been designated as finalists. Final Application Review and Winner Selection Organizations selected for investment will be notified in November. We will ask organizations selected for investment to answer a final set of questions that enable us to perform a more detailed due diligence review. For-profit grantees or other entities subject to expenditure responsibility will have to answer a specific set of questions to enable the Foundation to comply with its IRS obligations as a private foundation. For more on expenditure responsibility, please review the ER Guidelines in the application portal. Upon successful completion of this review, finalists will be asked to sign grant agreements with the Foundation. Execution of Grant Terms and Conditions All grant awards will be contingent on execution of a definitive grant agreement. The Representative Terms and Conditions can be found on the Application Portal. These Terms and Conditions have been developed specifically for this RFP and are not negotiable. You are advised to ensure that your institution can accept these Terms and Conditions at the time of proposal submission. Grantees subject to expenditure responsibility may be subject to additional terms and conditions If your proposal is selected, you will have a very limited amount of time after the notification of award to accept the grant and return the award letter with an appropriate institutional signature. You must return a fully executed Grant Agreement to the foundation post-marked no later than the date indicated by the Grants Management team to receive a grant award. Intent and Disclaimer This RFP is made with the intent to identify organizations to build solutions as described in this RFP. The Foundation will rely on an organization’s representations and consider them to be truthful as described. The Foundation assumes it can be confident in an applicant’s ability to deliver the activities described in this RFP. The responses will be incorporated into a future grant agreement should the Foundation wish to support the proposal submitted by the applicant. This RFP is not an offer to enter into a funding agreement. The Foundation assumes no responsibility for your cost to respond to this RFP. Until a written funding agreement is fully executed, the Foundation will have no obligations to any applicant. The Foundation has put in place policies and procedures to restrict public dissemination of grant application materials including, when possible, having external reviewers sign confidentiality agreements and requiring that reviewers destroy or return to the foundation all copies of information acquired or created during the course of performing a review. In some instances, we are unable to put in place confidentiality agreements or to police the use of grant application materials. As a general policy, the Foundation does not publicly disseminate or "publish" proposals or supporting information related to grant applications. For IRS compliance reasons, we are required to publish a list of grants that we have made. We also provide a general description of the grant on our web sites including www.gatesfoundation.org. These brief descriptions are also made available in press releases and other marketing materials. To identify and avert conflicts of interest among reviewers, reviewers will not be permitted to submit proposals, or to review proposals from organizations in which they have self-identified conflicts of interest. Page 13 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal More Information Reference Blank, Steve. The Startup Owner’s Manual: The Step-by-Step Guide for Building a Great Company. Pescadero: K&S Ranch, 2012. Blank, Steve. “Why the Lean Startup Changes Everything.” Harvard Business Review. May 2013. Chatterji, Aaron and Benjamin Jones. 2012. “Harnessing Technology to Improve K-12 Education.” Coalition for Evidence-Based Policy. 2012. “Rigorous Program Evaluations on a Budget.” Costa, Kristina. “Finding ‘What Works’ in Education.” Center for American Progress. 9 Feb. 2012. 15 July 2013. http://www.americanprogress.org/issues/open-government/news/2012/02/09/11053/finding-what-works-ineducation/. Education Market Research. 2012. The Complete K-12 Report®: Market Facts & Segment Analyses . Fullan, Michael and Katelyn Donnelly. 2013. “Alive in the Swamp: Assessing Digital Innovations in Education.” Gage, Deborah. “The Venture Capital Secret: 3 out of 4 Startups Fail.” The Wall Street Journal Online. 19 Sep. 2012. 16 July 2013. http://online.wsj.com/article/SB10000872396390443720204578004980476429190.html Global Silicon Valley Advisors. 2012. “Fall of the Wall: Capital Flows to Education Innovation.” Gray, L., Thomas, N., & Lewis, L. 2009. Teachers' Use of Educational Technology in U.S. Public Schools. Washington, D.C.: U.S. Department of Education. Havens, David. 2012. “A Closer Look at K12 EdTech Venture Funding.” Innovation and Entrepreneurship in Education – NewSchools Venture Fund Blog. 20 Dec. 2012. 15 July 2013. http://www.newschools.org/blog/closer-look IDEO & The Bill and Melinda Gates Foundation. 2013. “Catalyzing Ed Tech: User Journeys.” Lee, Mike. “The Top EdTech Frustrations of Educators.” EdShelf Blog. 9 Apr. 2013. 15 July 2013. http://edshelf.tumblr.com/post/47541656401/the-top-edtech-frustrations-of-educators Maxwell, Lesli A. “Study: Hybrid Algebra Program ‘Nearly Doubled’ Math Learning. Education Week. 20 June 2013. 15, July 2013. http://blogs.edweek.org/edweek/inside-schoolresearch/2013/06/government_study_finds_gains_f.html?cmp=ENL-EU-MOSTPOP mind the gap & The Bill and Melinda Gates Foundation. 2013. “Teachers and Technology.” Purcell, Kristen and Alan Heaps. 2013. “How Teachers are Using Technology at Home and in their Classrooms.” Ries, Eric. “Minimum Viable Product: A Guide.” Startup Lessons Learned. 3 Aug. 2009. 15 July 2013. http://www.startuplessonslearned.com/2009/08/minimum-viable-product-guide.html. Rogers, Everett M. Diffusion of Innovations. 5th ed. New York: Free Press, 2003. The Bill and Melinda Gates Foundation. 2012. “Innovation in Education: Technology and Effective Teaching.” US Department of Education Office of Educational Technology. 2013. “Expanding Evidence Approaches for Learning in a Digital World.” About the Bill & Melinda Gates Foundation Page 14 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal Guided by the belief that every life has equal value, the Bill & Melinda Gates Foundation works to help all people lead healthy, productive lives. We work with partner organizations worldwide to tackle critical problems in four program areas. Our Global Development Division works to help the world’s poorest people lift themselves out of hunger and poverty. Our Global Health Division aims to harness advances in science and technology to save lives in developing countries. Our United States Division works to improve U.S. high school and postsecondary education and support vulnerable children and families in Washington State. And our Global Policy & Advocacy Division seeks to build strategic relationships and promote policies that will help advance our work. Our approach to grant making emphasizes collaboration, innovation, risk-taking, and, most importantly, results. To learn more about the foundation's work, visit www.gatesfoundation.org. Page 15 of 15 © 2013 Bill & Melinda Gates Foundation Request for Proposal