Evaluation Criteria FITT – Fostering Interregional Exchange in ICT Technology Transfer – www.FITT-for-Innovation.eu Except where otherwise noted, this work is licensed under a Creative Commons Attribution 3.0 License. Criteria for evaluation of transfer projects • This practice is designed to assist in the preliminary assessment of researchgrounded technology projects for their commercialization potential in the realm of technology transfer. • The process of assessing research projects is necessitated by the high failure rate, and resulting high cost, of technologies either prior to reaching the market or once in the market. • The Evaluation Criteria are intended to provide guidance for assessing an idea, a technology or a research project, at an early-stage of technology transfer (thus prior to product development). 2 | March 2011 Evaluation criteria The evaluation process • Project evaluation may take place at various stages Early-stage (proof-of-concept “maturation” towards technology transfer) Pre-incubation Incubation • Our focus is Early Stage Project Evaluation, which may appear In a continuous manner (or at regular intervals) Based on a CFP (Call For Proposal, typically once per year) • Such early stage evaluation covers : Evaluation criteria A process for the application of these criteria, including the structure/organization of the evaluation committee The current practice focuses on recommended Evaluation Criteria 3 | March 2011 Evaluation criteria Illustration of the evaluation process Early-stage eval Research Development Pre-incubation entry evaluation Incubation entry evaluation Proof-of-concept Process : Evaluation criteria Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd 4 | March 2011 Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa • Description of project to be evaluated (document) Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee EeeeeBbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Aaaaaaaaa Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Eeeee Eeeee Aaaaaaaaa Bbbbbbbbbb Bbbbbbbbbb Cccccc Ddddd Eeeee Eeeee • Evaluation criteria • Jury (evaluation committee) Evaluation criteria Market Coverage/definition • Evaluation criteria should cover three main aspects of a project: Technical aspects Market assessment Team considerations • They should be defined and published in advance in order to allow the evaluated teams to adapt to the process • They will be used to establish the overall process, evaluation documents and the selection committee • Evaluation criteria may be used by the evaluation committee to Allocate funds/resources to selected projects Provide consultancy to the project team (for example, to coach the team on aspects considered as “weak”) 5 | March 2011 Evaluation criteria Evaluation criteria Possible evaluation criteria • Lots of possible evaluation methods/criteria are mentioned in the literature • Several possible groups of criteria : Originality of the innovation Scientific return/opportunities laboratory Project feasibility Profile of the inventor for the Business opportunity Positive ROI/NPV estimations Venture value Market opportunities/threats Regulatory constraints Potential users IP (protection issues, prior art) Business model Scientific relevance of the project Lab support Financial return Team aspects Realism of the announced plan Social & economical impact Risk management Potential applications Production issues 6 | March 2011 Evaluation criteria Focus on first-stage evaluation criteria Most important criteria for first-stage evaluation Originality of the innovation Scientific return/opportunities for the laboratory Project feasibility Potential users Scientific relevance of the project Team aspects Risk management Profile of the inventor Business opportunity Positive ROI/NPV calculations Market opportunities/threats IP (protection issues, prior art) Lab support Regulatory constraints Realism of the announced plan Potential applications Social & economical impact Venture value Business model Financial return Production issues Deemed premature for the evaluation early-stage technology transfer projects (prior to product development). 7 | March 2011 Evaluation criteria The DIGITEO example - Global positioning The OMTE checklist is used for maturation projects : 8 | March 2011 Evaluation criteria Timing of the annual call for proposal → March : launch of call for proposal/deadline for submissions Long selection process → April : preselection of 10 projects → May: coaching by Digiteo’s marketing team → June/July : final application, oral présentation, deliberation, final decision → September : launch of proof-of-concept activities for the projects selected Process of Digiteo’s maturation programme 9 | March 2011 Evaluation criteria From proposal to selection • ~ 10 proposals per year • Preselection classification performed by Digiteo’s scientific committee and marketing staff • Coaching : work on the three components technology/marketing/IP of the proposal and submit presentation for the final selection • Selection process : External experts (technology transfer specialists from : industry cluster, incubator, Paris region, OSEO innovation fund, chamber of commerce, etc.) Digiteo’s technology transfer committee Formal selection announced by Digiteo’s steering committee • Average of 5 projects selected per year 10 | March 2011 Evaluation criteria Selection steps 4. Final decision DIGITEO’s Steering Committee Scientific Committee 1. Scientific relevance Technical differentiation 11 | March 2011 Expert Panel 2. TT potential Value creation Evaluation criteria Technology Transfer Committee 3. Recommandations Digiteo’s evaluation checklist 12 | March 2011 Evaluation criteria DIGITEO – Method/criteria • Evaluation criteria used for the OMTE call for projects : « Product/technology » aspects Originality/uniqueness and scientific relevance, project feasibility and opportunities created for the laboratory. « Market » aspects Ongoing research contracts and IP related to the project, first applications and users considered. « Team » aspects Support of the laboratories in the process, project manager identified to manage the project, realism of the planning proposed and evaluation of the risks by the applicants. • Method : Evaluation of the applications according to the 12 criteria Evaluators apply assessment scores from 1 to 3 (3 being the highest) 13 | March 2011 Evaluation criteria DIGITEO – « Product » criteria 1. Originality of the innovation • Originality/uniqueness in comparison with state-of-the-art ? • Definition of the future « product » ? • Positioning compared to competitors ? 2. Scientific relevance of the project • Compatibility with the research themes covered by Digiteo ? • Scientific excellence in the field? • Degree of scientific maturation ( is the technology close to a « product ») ? 3. Project Feasibility • Technical feasibility of the project? • Feasibility of the planning, with regard to a transfer? • Description of the transfer model envisaged (transfer to an industrial partner / creation of start-up) ? 4. Scientific opportunities created for the laboratory • Consequences of the development on the scientific activities of the lab ? • Future impact of the project on the lab’s strategy ? • Impact on the external communications of the lab? 14 | March 2011 Evaluation criteria DIGITEO – « Market » criteria 5. Ongoing research contracts • Ongoing contracts with industrial partners? • Other contracts/ scientific activities ? • Since when? For how long? 6. Intellectual property (patents, know-how) • Background knowledge of the teams involved ? • Protection envisaged (foreground) for the new knowledge and software derivating from it; • Is an IP analysis requested by the teams (analysis of the prior art, patent landscape and « freedom to operate ») ? 7. First potential applications • Types/examples of applications ? • Value Proposition (solution to which problem) ? • Applications realised by which kind of company (software company, service provider) ? 8. First potential users • Existing and potential actors/ partners to target for the transfer? • Example of end-user for the integrated solution ? • Draft definition of the targeted market (size, segmentation, competitors) ? 15 | March 2011 Evaluation criteria DIGITEO – « Team » criteria 9. Support of the laboratories • Support of the laboratories involved ? • Balance between the teams involved in the project (complementarity, synergy) ? • Commitment to a real transfer ? 10. Project manager in charge • Profile of the project manager and implication in the project ? • Capacity of managing all aspects of the project, keeping with the transfer objective? • Motivation to handle the 3 aspects : technical, IP, marketing ? 11. Realism of the planning Realism of the planning with regards to the 3 aspects: • Technical • IP • Marketing 12. Evaluation/ consideration of the risks Identification and management of the risks : • Technical • IP • Marketing 16 | March 2011 Evaluation criteria DIGITEO - Assessment • Useful tool to be used as a checklist throughout the evaluation process • The final selection has to include the assessment of the presentation made in front of the jury. Grade given by the jury is based for 50% on written application and 50% on the oral presentation. • The jury should include a majority of external experts • Final selection : classification/ranking of the presented projects (top 5 selected) • Some « Digiteo specifics » not to be considered for a generic checklist 17 | March 2011 Evaluation criteria Pros & Cons PROs CONs • This practice attempts to formalize methods that are already in use (most of the time on an ad hoc basis) • Only a selected number of criteria are highlighted • Some criteria may need to • The methodology and associated tools (call for proposal, criteria, etc.) are readily available and can be adapted to each individual case 18 | March 2011 Evaluation criteria be further developed Rationale • Methodology developed by Digiteo in order to manage the incoming flow of technology transfer proposals • Need for a consistent set of criteria for all steps of evaluation process, communicated transparently to all involved partners : project teams, internal Digiteo evaluators, “technology transfer coaches” and external experts • Without this methodology, involved parties would get the impression that projects might be evaluated/selected based on obscure reasons. This would leave the doors open for debate, accusations for “unfair competition” and backstage lobbying 19 | March 2011 Evaluation criteria Outcome • The approach turned out as expected Final selection (with external experts) is based on relative ranking among the presented projects The scoring system is only used for individual evaluation purposes • Coaching the teams before the final applications, supported by the evaluation checklist, proved to have a great impact on the oral presentations. Final applications show significant improvements (in terms of technology transfer readiness) compared with initial applications. • Feedbacks indicate that Digiteo community (researchers, external experts) judge the approach fair and clearly communicated 20 | March 2011 Evaluation criteria Lessons learned • A strict, fixed procedure can sometimes be counter-productive. For this reason, scoring was not made compulsory for the final evaluation and used as a guideline during the selection committee. • Special attention must be paid to the management of unsuccessful project teams. A debriefing session is organised with each eliminated team in order to: • debrief the teams that were not selected • clearly communicate the reasons for not being selected • focus on things to be improved (and how to improve them) • encourage them to apply again with an enhanced proposal 21 | March 2011 Evaluation criteria Plans for the future The approach should be further developed/detailed : Definition of terms Explanation on how to apply each of the listed criteria (with some examples) 22 | March 2011 Evaluation criteria Evaluation Criteria in practice with IBBT IBBT wanted to use evaluation criteria for the following goals: scoring of new proposals at intake scoring & ranking of proposals within the same call quick visual overview of the scoring between several proposals evaluation or follow-up of a Bootcamp project or incubation project The approach: master list made up of all applicable evaluation criteria. From this master list , subset of criteria for each type of project was defined Validation of those criteria sets by means of a proof of concept in excel format on different kind of projects 23 | March 2011 Evaluation criteria Lessons learned Using evaluation criteria to objectify and structure the assessment of a project based on different domains, different criteria and sub-criteria proved to be very helpful: • in panel discussions: as a common base for discussion between the members of an expert panel to select the most promising business ideas to go through to the coaching track of iBoot • in coaching trajectories: as feedback tool to the team indicating weak and strong points and getting constructive feedback of experts • in follow-up meetings: as a common base for discussion between the innovation board members about the progress of incubation projects Since assessments are made by different coaches, innovation board members and experts, we added to the scoring of 1-4 some illustrative “quotes” . This makes the nuances between scoring-numbers more justifiable. Example – Criterion - Suitability of the approach • Score 1: There is no connection whatsoever between the approach and the intended results • Score 2: The proposed approach is not very consistent with the intended innovation purpose • Score 3: The approach is consistent with the innovation purpose • Score 4: The approach is very well targeted 24 | March 2011 Evaluation criteria Evaluation Criteria in practice with ICBS At ICBS, the evaluation criteria of Digiteo have been used for the evaluation of the final projects of the IED programme. The IE&D program (Innovation, Entrepreneurship & Design) is part of an MBA course. Students embark on a live entrepreneurial journey and investigate the commercial potential of their own idea or one being developed by Imperial Innovations or Innovations Royal College of Art Modification of the practice needed before testing: New criteria are introduced: - Considerations of financials (economic feasibility) - Scalability (production & operation) 25 | March 2011 Evaluation criteria Lessons learned • Qualitative approach of the indicators, making it easy to quickly evaluate the projects. • Useful to inform the candidates of the criteria that will be used by the jurors before the evaluation. • Good tool to start a systematic evaluation of projects. • Suggestion to add a general evaluation of the opportunity, as a preliminary step in the process. Analyze if there is a real problem (demand pull) or if it a technology push type of opportunity. • Suggestion to include indicators on the economic feasibility of the project and on production/scalability. The goal is to see if the predicted production process/time is reasonable. The scalability can be evaluated by taking into account the time and money invested for the prototype. As a consequence, some of the items may be combined in order to introduce these new concepts. • Re-writing some of the items would make them easier to be understood without the guideline (the list of concepts). 26 | March 2011 Evaluation criteria Evaluation Criteria in practice with CRP Henri Tudor The project definition process within Tudor has been reviewed. The goal was to organize the earliest phase of project development and include a screening phase where the project’s idea is assessed, including criteria on technology transfer. Difference of context : criteria for pre-projects, where the project’s idea is developed, formalized and validated. Objective •Create awareness for Technology Transfer in researcher’s mind when they write their project’s proposal •Improve assessment of project’s idea and have a referential for the entire organisation •Have a better follow-up project to foster maturation project and technology transfer 27 | March 2011 Evaluation criteria Lessons learned • Putting in place a screening process has created some fears within the researcher community, including fear of breaking creativity. But finally, screening meetings are now a place of discussion where the project proposal can be assessed and enriched. The fact that these criteria are already used by peers is also a guarantee for the management of the organisation and researchers. • Important to spend time on change management, when implementing a new process that will have a strong impact on the organisation. • The outcome for Tudor is a better overview on projects developed within the organisation for all stakeholders and integration of the transfer perspective at the very beginning of projects. 28 | March 2011 Evaluation criteria Suggested Readings Link to code book assessment; method; process; selection Link to related websites OMTE call for proposal and projects selected during previous editions: http://www.digiteo.fr/Digiteo_OMTE 29 | March 2011 Evaluation criteria