POST PRINT VERSION. Accepted by International Journal of Educational Research on 27 August 2012. *Note – this is a copy of the final draft version submitted on 4 July 2012 after peer review. International Journal of Educational Research 56 (2012) 23–34 http://dx.doi.org/10.1016/j.ijer.2012.08.001 Author details: Corresponding author: Adrian Cherney *Dr Adrian Cherney School of Social Science The University of Queensland Brisbane, St Lucia 4072 ph + 61 7 3365 3236 fax + 61 7 3365 1544 email: a.cherney@uq.edu.au Dr Jenny Povey Professor Brian Head Professor Paul Boreham Michele Ferguson Institute for Social Science Research The University of Queensland Brisbane, St Lucia. Acknowledgements: This project is supported through ARC Linkage project: LP100100380. 1 What influences the utilisation of educational research by policy-makers and practitioners? – The perspectives of academic educational researchers Abstract In the field of education much has been made of the need for academics to engage more closely with policy-makers and practitioners in the process of knowledge production and research uptake. This paper reports results from a survey of academic educational researchers in Australia on their experience of research uptake and engagement with policy-makers and practitioners. We examine a range of variables to understand factors influencing the use of educational research. The results indicate that while research uptake is enhanced through mechanisms that improve the intensity of interactions between academics and end-users, the dynamics of research collaborations have a significant bearing on research use. Our findings provide insights into the challenges that can be confronted when academics engage in research aimed at influencing policy or practice. Keywords: research utilisation, research collaborations, education, knowledge translation, policy impact, academic research 2 1: Introduction In the field of education the observation has been made that academic research rarely has a policy impact and often fails to meet the needs of policy-makers and practitioners (Coburn & Talbert, 2006; Hemsley-Brown & Sharp, 2003; Hess, 2008; Hillage et al 1998; Levin & Edelstein, 2010; Oancea, 2005). This disjunction is partly seen as originating in communication problems between policymakers, practitioners and academic researchers, drawing on the argument that they live in different worlds with differing languages, values and professional rewards (Bell et al., 2010; Kirst, 2000; Levin, 2011; Orland, 2009; Vanderlinde & van Braak, 2010). There is some validity to this argument, with studies in the field of education indicating that educational researchers, bureaucrats and teachers often have different priorities and perceptions about what constitutes useful and valid research, the role of theory, data quality and research methods, project outcomes, brevity of results and the practicality of research recommendations (Bell et al., 2010; Coburn & Talbert, 2006; Cousins & Leithwood, 1993; Honig & Coburn, 2008; Levin & Edelstein, 2010; Saha, Biddle, & Anderson, 1995; Vanderlinde & van Braak, 2010; Wilkins, 1988; Zeulie, 1994). Greater collaboration between academic research producers and users, or consumers of research, is seen as one way of addressing the dissonance between knowledge production and its transfer or translation to policy and practitioner contexts. There is some evidence that when academic researchers and policy-makers or practitioners work closely in the formulation and execution of research projects, the research is more likely to have an influence on policy or practice (Cordingley, 2008; Cousins et al., 1996; Cousins & Simon, 1996; Huberman, 1990; Nutley et al., 2007). However, closer collaboration alone is insufficient to ensure that research has a policy or practice impact, with studies demonstrating that a range of variables influence the uptake and use of academic social research by nonacademic end-users (Bell et al 2010; Bogenschneider & Corbett, 2010; Cherney & McGee, 3 2011; Huberman, 1990; Landry et al., 2001a, 2001b; Weiss & Bucuvalas, 1980). Moreover, research collaborations can be inherently problematic with participants’ involvement influenced by individual and institutional constraints and contingencies (Bell et al 2010; Bogenschneider & Corbett, 2010; Coburn & Talbert, 2006; Cousins & Simon, 1996; Edwards et al 2007). The issue of closer synergies between educational researchers and non-academic endusers raise important issues concerning the role that educational research should play in relation to policy and practice. As Hammersley (2007) points out, ones position on this issue is influenced by judgments on whether academic educational research should be integral to practice or ought be judged as a value in its own right. While more nuanced positions are often adopted that recognise the multi-dimensional value of educational research (e.g. see Cooper, Levin and Campbell 2009) this is a highly contested issue (Burkhardt & Schoenfeld, 2003; Lingard, 2011). The intensity of debate about the value of academic research has also been increased through University Research Assessment Exercises, such as the Research Excellence Framework in the UK, which has required academics to demonstrate the impact of their research. Academic researchers in Australia are subject to similar pressures (e.g. through the Excellence in Research Australia initiative and at the time of writing the Excellence in Innovation for Australia trial). While our aim here is not to elaborate upon all these issues, our paper does provide new analyses relevant to the debate around research impact and evidence-based policy in the field of education. Using survey data from academic university researchers in Australia who engage in research collaborations with external partners, the paper principally aims to examine factors that influence the uptake of social research, as interpreted through the experience of knowledge producers in the field of education. Using the scale of research 4 utilisation (Knott & Wildavsky, 1980) we examine factors that appear to influence reported levels of research impact. The paper is organised as follows. Firstly, the explanatory model underpinning this study (i.e. the scale or ladder of utilisation) will be discussed. Secondly, the paper outlines the data collection methods used for the survey administered to Australian academics. Thirdly, key results from the sample of educational researchers are provided, focusing on reported levels of research utilisation and variables that appear to influence knowledge transfer and application. Finally, the paper discusses the results (and some data limitations) and concludes with broader observations about the study of research uptake in the field of education. 2: Literature review 2.1: Measuring Research Use When it comes to measuring research utilisation, no single conceptual model has been unanimously adopted (Belkhodja et al., 2007; Lester, 1993). One reason for this is the methodological problem associated with specifying the dependant variable of research use given it can be defined either as a process or an outcome. Furthermore social research can provide answers to technical questions, such as “did this program work?” to helping policymakers or practitioners interpret problems in ways that change their understanding about issues or choices (Biesta 2007; Nutley et al., 2007). Hence the complexity surrounding the use of academic social research can make it difficult to measure, particularly when attempting to quantify its interruptive function (Beista 2007). Scales of research use can be particularly helpful in understanding that the breadth of social research usages ranging from practices that encompass transmission through to actual application (Cherney & McGee, 2011). Such scales can be valuable because they can be used to identify how utilisation is related to various decision-making processes, particularly concerning the actions of the producers and consumers of social research (Landry et al., 2001b). 5 Methodologically this study replicated a modified version of the Knott & Wildavsky (1980) research use (RU) scale, , similar to that adopted in the study by Landry et al (2001a). This scale was adopted because it has been frequently cited in the literature, has been used to measure research use among government officials and academics, and has been shown to be reliable (Landry et al., 2001a; Lester and Wilds 1990; Lester 1993)1. Conceptually the research use scale can be referred to as the “ladder of utilisation” and table 1 provides the descriptions for each stage of research use (or rung of the ladder), as presented in our questionnaire to Australian social scientists. The benefit of this scale is that it operationalises research use as a cumulative process that progresses through a number of stages: transmission, cognition, reference, effort, influence and application. The scale is cumulative in the sense that cognition builds on transmission, reference builds on cognition, effort on reference, influence on effort, and application on influence. The RU scale has been criticised as perpetuating a linear understanding of research utilization (Davies & Nutley, 2008). However it does recognise the fact that the research utilisation process varies between a range of activities that involve knowledge transfer and uptake (Cherney & McGee, 2011; Knott & Wildavsky, 1980; Lester, 1993). <Insert Table 1> 2.2: Independent Variables Influencing Research Use Just as there is no agreed conceptual model relating to research utilisation, there is no definitive list of variables developed to help predict knowledge use (Lester, 1993). Most studies have categorised variables under broad groups relating to supply-side and demandpull factors, as well as dissemination and interaction variables. Supply-side factors include research outputs and the context in which the researcher works. These can include the types 1 Another key reason this scale was chosen was to allow for international comparison with studies such as Landry, Amara and Lamari (2001a) who also adopted the research use scale. We have not provided such comparative results because it falls outside the scope of this paper. 6 of research outputs produced by academics (e.g. qualitative or quantitative studies2), whether research is focused on non-academic users, the importance of internal or external funding sources, and the institutional drivers that influence the initiation of collaborations with external partners and end-users (Bogenschneider & Corbett 2010; Cherney et al., 2011). Demand-pull factors concern whether end-users consider research to be pertinent, whether it coincides with end-users’ needs, whether users accord it credibility, and whether it reaches users at the right time to influence decision-making. Added to this are organisational processes such as level of skills to apply research knowledge, that could inhibit uptake of research and thus influence the overall demand for academic research within end-user organisations (Belkhodja et al., 2007; Coburn & Talbert, 2006; Ouimet et al 2009). Dissemination variables relate to efforts to adapt and tailor research products (e.g. reports) for end-users and to develop strategies focused on the communication of research (Huberman, 1990). The assumption is that the more researchers invest in adaptation and dissemination, the more likely research-based knowledge will be adopted. Adaptation includes efforts to make reports more readable and easier to understand, efforts to make conclusions and recommendations more specific or more operational, efforts to focus on variables amenable to interventions by users, and efforts to make reports appealing (Cherney & McGee, 2011). Dissemination efforts include strategies aimed at communicating research to targeted endusers, such as when researchers use different social media to communicate their research messages, hold meetings to discuss the scope and results of their projects with specific users or partners, and target particular forums, e.g. reporting on their research to government committees. Finally, interaction variables focus on the intensity of the relationships between knowledge producers and potential users. The types of factors considered relevant include 2 Both qualitative and quantitative approaches can serve different purposes, relating to the technical and interruptive functions of social research (see section 2.1). We have used this distinction between qualitative and quantitative approaches in order to be able to measure and identify whether chosen research methodology mediates the impact of social research. 7 informal personal contacts, participation in committees or experience with research partnerships, e.g. the number of research partnerships an academic has engaged in (Huberman, 1990; Landry et al., 2001a; Lomas, 2000). 3: Current Study The data used in this research were drawn from a broader study examining evidence-based policy and practice (Cherney et al., 2011). The study has four phases and the data reported here was obtained from Phase 1, which used a purposive sampling technique to target academic social scientists in Australian Universities3. The final sample recruited was 693, which constitutes an overall response rate of 32 per cent. For the purpose of this analysis, only data pertaining to academics who identified their primary research discipline as education have been used (n=156). The academic survey was partially based on existing items or scales (Bogenschneider & Corbett, 2010; Landry et al., 2001a, 2001b) but with additional items included to gauge the dynamics of research partnerships. 3.1: Dependent variable Knowledge utilisation was measured using a validated version of the Knott and Wildavsky (1980) research use scale. As indicated the scale is based on six stages namely: transmission, cognition, reference, effort, influence, and application. For each of these six stages respondents were asked to estimate what had become of their research using a 5-point scale ranging from 1 (never), 2 (rarely), 3 (sometimes), 4 (usually), to 5 (always). 3 This involved targeting Fellows of the Academy of the Social Sciences in Australia (ASSA) as well as Australian academics who had secured at least one Australian Research Council (ARC) grant between 2001 and 2010 within the field of social and behavioural science. Academics who met either of these criteria were sent an electronic survey to complete. ASSA fellows are recognised for their outstanding contributions to the social sciences in Australia and abroad and dominantly comprise senior academics. Australian Research Council (ARC) grants are national competitive grants and funds a significant proportion of research activity in Australian Universities. The reason this form of sampling was adopted was because studies have shown that seniority and the number of external competitive research grants are key determinants of engagement with nonacademic end-users (Cherney & McGee 2011; Landry, Amara, and Lamari 2001a, 2001b). 8 Previous researchers (Cherney & McGee, 2011; Landry et al., 2001a) have used this scale cumulatively (with each stage building upon the next) and assigned a value of 1 when respondents replied always, usually, or sometimes, and with all other responses assigned the value of 0 or a fail. There are two ways that this cumulative approach can be analysed. The first is to run a separate logistic regression for each stage of research utilisation as Landry et al (2001a) did in their study. Hence respondents who pass all six stages would be represented in each stage or regression model (see Figure 1). This is particularly problematic with our sample, because the majority (75%) of the sample reported they passed all six stages. Hence the question arises whether such a method would really be determining what predicts movement from one stage to the next, or whether progression across each stage is masked by the dominant group. In order to address this criticism, a second approach would be to create an ordinal variable with seven levels, including in each level only those individuals who passed that level. Thus respondents in each level would be unique. Table 2 presents the number of respondents categorised in each level/echelon according to such progression criteria. For instance, two percent of the sample passed the transmission stage but did not progress further. However, an ordinal logistic regression analysis is not possible due to our small sample size and the number of cases in each level. The next possible option would be to examine whether these stages are in fact exclusive. Does failure in one stage preclude academic researchers from progressing to other stages? Or should these stages comprise an index? Descriptive statistics, as presented in Figure 2, illustrate that failure in one stage does not preclude academic researchers from passing subsequent stages. This is an important consideration because the process of research utilisation has been argued to be non-linear, with data in Figure 2 indicating that one does not necessarily have to traverse in sequence each rung of the research utilisation ladder to reach the ultimate stage - i.e. application. A factor analysis of the items (or stages), revealed a 1- 9 factor solution and a Cronbach’s alpha coefficient of 0.77 (Table 3). Thus, the results indicate that these items are measuring one construct and that the index seems to be reliable. It was thus decided to use the items as an index to measure research use. A mean index score was calculated for all 6 six stages. The mean score for the research utilisation index is presented in Table 3. <Insert Figure 1> <Insert Table 2> <Insert Figure 2> 3.2: Independent variables A number of indices were created and included in our model as independent variables. The items used in each index were determined by factor analyses, with each index comprising a 1factor solution. The Cronbach’s alpha coefficients for these independent variables are presented in Table 3 and detailed descriptions of index compositions are presented in Appendix 1. <Insert Table 3> Descriptive statistics for each independent variable are presented in Table 4. Academic researchers from the discipline of education indicated that academic funding (i.e. national competitive grants such as Australian Research Council grants, and internal University funds) were more important than funding from government and non-government agencies in ensuring their research is conducted. Academic researchers indicated that the ‘relevance’ of the research is given a higher priority by end-users, compared to other features such as the quality or feasibility of the research. A higher level of importance was attributed to tailoring research to meet the needs of end-users and meetings to discuss findings with end-users. Table 4 also illustrates a high level of agreement among academic educational researchers concerning the fact that they encounter barriers in the transfer and uptake of their research. A 10 very high level of importance is accorded by academic educational researchers to the use of refereed publications as a method through which to disseminate their research. The number of research partners with whom researchers engaged ranged between 0 and 35, with an average of 6 research partners per researcher. The number of grants received by these academic educational researchers varies between 0 and 44, with the average researcher having received 8 grants. In general, our sample has a high level of experience in engaging in research partnerships and securing research grants. <Insert Table 4> 3.3: Data analysis Given that our dependent variable is approximately continuous, an Ordinary Least Squares (OLS) regression model was used to estimate the associations between research utilization (our dependent variable) and a number of explanatory variables such as benefits and barriers associated with engaging in research with policy-makers and practitioners. As a preliminary check, we examined the correlations between all variables in the model. They ranged between .002 and .68, suggesting that multicollinearity was unlikely to be a problem (the correlation matrix was too large to depict in the Appendix). This was confirmed by a relatively low value of the mean Variance Inflation Factor (VIF) of 1.66, with the individual variable’s VIFs ranging from 1.18 to 2.48. The four highest correlations were between problems relating to the orientation of research partnerships and ‘consequences’ of investing in research partnerships (0.68); ‘consequences’ of investing in research partnerships and barriers academics experience in the transfer and uptake of research by end-users (0.57); importance of using contacts, seminars and reports to present research to policy-makers and practitioners and importance of meetings and dissemination activities with end-users (0.55); and importance of meetings and dissemination activities with end-users and importance of tailoring research when end-users are the focus (0.55). All four correlations were statistically 11 significant. 3.4: Regression Results The regression results are presented in Table 5. The results indicate that eight variables were significantly related to the utilisation of educational research. The more academic researchers perceived collaboration with external partners as beneficial, the more likely they report utilisation. As the number of grants increases so does the likelihood of research utilisation. The more negative the perceived consequences for academic educational researchers when engaging in research partnerships, the less likely it is reported that their research will be utilised by policy-makers or practitioners. The importance of tailoring the research for endusers is positively and significantly associated with reported levels of research use. Academic educational researchers also reported that when end-users felt that research was relevant it was more likely to lead to utilisation. Finally academic researchers indicated that the more policy-makers or practitioners prioritise the ‘feasibility’ of research (i.e. policy-makers or practitioners place greater emphasis on research being economically and politically feasible) the less likely were academics to perceive that end-users would use academic social research. When academic researchers perceived there to be problems associated with research partnerships they were less likely to report research uptake by external agencies. <Insert Table 5> 4: Discussion Before proceeding to a discussion of possible practical and theoretical explanations for this pattern of results, the methodological limitations of the present study should be noted. Firstly, the survey data is based on self-reports, which can be subject to social desirability biases. In the context of academic researchers reporting levels of research use this can be a salient factor in influencing responses to survey items (Davies & Nutley, 2008; Levin, 2011). Furthermore it is possible there is a self-referential process occurring in that respondent’s 12 reported level of research utilisation is reinforced by how relevant they perceive their research to be for policy-makers or practitioners and vice versa. We have not examined the reference point (e.g. specific project contexts) establishing why respondents believed their research had an impact or why problems were encountered. One additional way to measure judgments about research utilisation among academics is to adopt a two-step approach using a yes-no question, first about a particular research use stage (e.g. my research reports have been read and understood by end-users) and then asking respondents to indicate how frequently they think it occurs. This can help increase measurement accuracy. There is little doubt though that tracing the impact of research evidence and collaborations needs to be complemented by more detailed qualitative work. The survey data only gives insight into broad patterns of utilisation as reported by our academic sample – hence providing one possible perspective and potentially reflecting particular biases. This does not undermine the validity of the results reported here, because a full understanding of the process of research transfer and uptake requires a close examination of the experiences and perspectives of both knowledge producers and users, and an exploration of broad general patterns as well as specific instances of research uptake. The results reported above indicate that our sample of educational researchers recognise that, in order for their research to have an impact, there is a requirement to directly engage with end-users through meetings and dissemination processes and that it is important to tailor research projects and findings to end-user needs. Despite reports that academic researchers do not understand the needs of policy-makers or practitioners (Burkhardt & Schoenfeld, 2003), our sample were aware that non-academic end-users do have different priorities and perceptions when it comes to judging the relevance and use of research evidence. This awareness is perhaps linked to the fact that our sample was experienced in engaging with multiple research partners, making them mindful of these contextual issues. 13 Reported levels of research use was also shown to be heavily influenced by access to funding particularly national competitive funding and internal University funds, compared to others sources available from government and non-government agencies. Intuitively one would assume that this finding would be the other way around, given the applied focus of many government and non-government funding schemes. One explanation for this result is that a high reliance on academic funding sources is possibly related to the fact that government and non-government schools and educational departments face significant funding constraints and hence do not have large amounts of discretionary funds to direct towards research. Hence educational researchers in Australia need to rely heavily on national competitive academic research funds (such as the Australian Research Council) and internal sources within Universities (Holbrook et al., 2000). Such funding issues are not unique to the Australian context (Burkhardt & Schoenfeld, 2003). When it comes to the levels of research utilisation reported by our respondents, the perceived costs and benefits of engaging in research partnerships had a significantly strong relationship to the reported transfer and uptake of academic research. While much is made of the need for academics to work more closely in partnership with external agencies, the potential consequences and costs of doing so cannot be overlooked, especially in relation to problems that may arise in the context of managing research partnerships and the priorities of partners. Given the differing “world views” about research and the institutional cultures driving academics and policy-makers or practitioners, tensions in collaborations are inevitable (Bogenschneider & Corbett, 2010; Edwards et al., 2007). A major challenge for educational researchers is that developing the necessary know-how to manage collaborations more effectively is not part of mainstream academic training and is typically a skill developed over time through experience and mentoring. 14 Two noteworthy results that arose from the regression analysis are the findings about the reported importance of the ‘feasibility’ of research and the industry-orientation of partnerships in relation to perceived levels of research use. The ‘feasibility’ variable comprised three items: research recommendations are seen as economically and politically feasible and research findings support a current position or practice (see Appendix 1). When feasibility is seen as a strong priority, our respondents reported that academic research is less likely to be used. This finding points to a key lesson that the quality of research evidence is not the only priority potentially driving research use and nor is it the single most important factor in determining uptake. Even if one accepts the realities of end-user preferences and tries to work within the policy and practice space offered by collaborations with external endusers, the ‘orientation’ of research partnerships can further influence/undermine levels of uptake. This relates to the finding that partnership orientation can create problems for research impact. This ‘orientation’ variable comprised five items (see Appendix 1) relating to the fact that partnerships can become dominated by what industry funding partners desire, which can challenge the integrity and rigor of the academic knowledge production process. It is little wonder some of our sample were weary of engaging in research collaborations. The findings discussed above do not mean that educational researchers should avoid engagement in research collaborations with government or non-government partners. On the contrary, the implication is that educational researchers need to be mindful of the priorities driving external agencies when they seek out research and engage in collaboration. What are the implications of our findings for the impact of educational research on policy and practice? One is that educational researchers need to become adept at understanding the priorities of policy-makers and practitioners and craft knowledge translation activities, such as research presentations, so they more effectively address the knowledge needs of non-academic end-users. This does though raise the risk of academic 15 educational research simply being co-opted and directed by pragmatic concerns, which has implications for knowledge production by potentially deflecting attention away from areas of research that prove uncomfortable for policy-makers or practitioners. Translating educational research is a difficult balancing act between ensuring research is relevant to non-academic end-users, but in a way that does not compromise its scholarly integrity. The problem though is that knowledge translation activities are not well understood in the academic social sciences (Levin, 2011). The work involved in collaboration does bring benefits to academic researchers in the context of providing opportunities to influence policy or practice within educational settings such as schools. However, academics must learn to tolerate the costs of such engagement. The willingness to endure such costs can be heavily influenced by the institutional incentive systems within academia including whether collaboration is seen as contributing to valued forms of scholarship. Our data also indicate that when educational academic researchers seek to undertake ’research for policy‘, their work must inevitably contend and compete with a mix of political and economic priorities and values that influence policy and practitioner decision-making (Head, 2008; Lingard, 2011). 5: Conclusion The analysis presented in this paper provides insights into the types of challenges academics face in generating impact through research collaborations. It provides empirical backing to the broader debate and commentary within the field of education as to why there appears to be dissonance between the evidence produced by educational researchers, and its transfer and uptake within policy and practitioner contexts. Our data indicates how this research/policy gap can be reduced, but it also signals that doing so through partnership synergies between academic researchers and external agencies is not a straightforward solution. Understanding the dynamics of different research collaborations with government and non-government 16 agencies is essential in ensuring that they are a success, and these dynamics require further study. While our data only provide insights into the perspectives of academics, investigating their experiences of collaborations and perceptions of research impact is essential for identifying how these processes can be improved. 17 Appendix I: Independent variables measures Research Approach Quantitative studies The quantitative research approach is a single item variable that reflects how often researchers use a quantitative approach such as surveys research, statistical analysis, and GIS in their research. The results reported are the percentage of respondents who indicated always or usually. These responses were recorded as 1, while all other responses were recorded as 0. Qualitative studies The qualitative research approach is a single item variable that reflects how often researchers use a qualitative approach such as interviews, focus groups, ethnography, and observation in their research. The results reported are the percentage of respondents who indicated always or usually. These responses were recorded as 1, while all other responses were recorded as 0. Researchers’ Context Research targeted to user This Index measures whether the majority of research conducted by academics is directed at policy-makers and practitioners. This index is comprised of four dimensions that range on a 4-point scale, ranging 1 (never) to 4 (always). The four dimensions are: (1) policy makers within government; (2) practitioners/managers within the public sector; (3) practitioners/managers within the community sector; (4) practitioners/managers within the private sector. Importance of academic funding This Index measures how important academic type funding is in ensuring research is conducted. This index is comprised of two dimensions that range on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The two dimensions are: (1) my university’s internal research funds; (2) funding organisations such as ARC, NHMRC, CRC. Importance of other funding This Index measures how important other funding sources are in ensuring research is conducted. This index is comprised of five dimensions that range on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The five dimensions are: (1) not for profit organisations; (2) federal government agencies; (3) state government agencies; (4) local government agencies; and (5) private sector organisations. 18 Benefits of collaborative research This Index is based on academic perceptions of the benefits of carrying out research in collaboration with government, industry or community sector partners. This index is comprised of ten dimensions that range on a 6-point scale, ranging from 0 (not applicable), 1 (strongly disagree) to 5 (strongly agree). The ten dimensions are: (1) I have been able to use data that would otherwise be difficult to access; (2) Research partnerships have provided me with opportunities for my research to have an impact on policy and practice; (3) Research partnerships have helped to increase my industry contacts; (4) My industry contacts have helped with developing future research projects; (5) Research partnerships enable me to generate extra income for my work unit; (6) Such projects have provided me opportunities to commercialise research outcomes; (7) Research partnerships have helped me with career advancement; (8) Such projects have required me to be pragmatic and realistic in relation to research outcomes for industry partners (9) Research partnerships have enabled me to publish in a broad range of publication outlets (10) I find projects with external partners more satisfying than fundamental “blue sky” research. Barriers academics experience in transfer & uptake of research This index is based on the barriers academics experience in the transfer and uptake of their research. This index is comprised of five barriers that range on a 6-point scale, ranging from 0 (not applicable), 1 (strongly disagree) to 5 (strongly agree). The five dimensions are: (1) There are high costs (e.g. time and resources) in translating the results of research for policy-makers and practitioners; (2) There are insufficient forums and networks available for bringing together researchers and non-academic endusers of research; (3) Academic reward systems do not adequately recognise dissemination work to non-academic end-users; (4) The academic requirement to publish primarily in peer-reviewed journals inhibits a focus on policy and practitioner audiences; (5) Networks and partnerships that might support research uptake are often undermined by turnover of contact staff in public agencies. Consequences of investing into research partnerships This index is based on problems relating to investing time and resources and accommodating partnership work that academic researchers encounter when carrying out research with partners from government, industry or the community sector. This index is comprised of ten items that range on a 6-point scale, ranging from 0 (not applicable), 1 (strongly disagree) to 5 (strongly agree). The ten dimensions are: (1) There are inadequate university resources 19 to support research partnerships with end-users; (2) I find there are different research orientations between academics and external partners; (3) You need to invest a lot of time in coordinating the work between different partners; (4) Confidentiality requirements often restrict what you can report and publish; (5) You can lose ownership of intellectual property; (6) You are subject to delays that impede your ability to publish results in a timely manner; (7) I am under pressure from my work unit to undertake contract research to meet budget requirements (8) External partners do not appreciate the full costs of research; (9) The ethics process can be time consuming and cumbersome; (10) The complexity of contractual arrangements can lead to delays in commencing research. Problems relating to orientation of research partnership This index is based on problems relating to the priorities and expectations of the partners when carrying out research with partners from government, industry or the community sector. This index is comprised of five items that range on a 6-point scale, ranging from 0 (not applicable), 1 (strongly disagree) to 5 (strongly agree). The five dimensions are: (1) I find there is pressure to produce favourable results for partners; (2) I believe such projects overemphasise applied outcomes; (3) I do not feel comfortable working on projects carried out in collaboration with industry or government agencies; (4) I feel that industry partners place too much emphasis on specific deliverables; (5) I feel that there is too much pressure to meet deadlines. Research Time This is a dummy variable created from the question asking academics to indicate the nature of their position, either research and teaching or research only. The research only was used as the reference group. Number of external grants This index is the sum of all the research grants (i.e. ARC discovery, ARC linkage, other external competitive grants) academics have received. Users’ Context End-users prioritise high quality research This index is based on academic researcher’s perceptions of what research characteristics end-users prioritise when using academically produced social science research. This index is comprised of seven dimensions that range on a 5-point scale, ranging 1 (not a priority) to 5 (high priority). The seven dimensions are: (1) high quality research; (2) unbiased findings; (3) adds to theoretical knowledge; (4) statistical analysis is high 20 quality; (5) findings can be generalised; (6) offers a new way of thinking; and (7) reputation of researcher. End-users prioritise the useability of the research This index is based on academic researcher’s perceptions of what research characteristics end-users prioritise when using academically produced social science research. This index is comprised of four dimensions that range on a 5-point scale, ranging 1 (not a priority) to 5 (high priority). The four dimensions are: (1) findings available when decisions need to be made; (2) findings have direct implications for policy & practice; (3) findings written in a clear style; and (4) report has brief summary of findings. End-users prioritise the feasibility of the research This index is based on academic researcher’s perceptions of what research characteristics end-users prioritise when using academically produced social science research. This index is comprised of three dimensions that range on a 5-point scale, ranging 1 (not a priority) to 5 (high priority). The three dimensions are: (1) recommendations are economically feasible; (2) findings support a current position & practice; and (3) recommendations are politically feasible. Policy-makers have found my policy focussed research to be relevant This index is based on the experience of researchers who have had a policy focus in their research and whether policy-makers found their research to be relevant. This index is comprised of three dimensions that range on a 6-point scale, ranging from 0 (don’t know), 1 (strongly disagree) to 5 (strongly agree). The three dimensions are: (1) relevant to their needs and expectations; (2) valid and reliable; (3) trustworthy. Barriers policy-makers experience in transfer & uptake of research This index is based on the barriers policy-makers are perceived to experience in the transfer and uptake of academic research. This index is comprised of two barriers that range on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The five dimensions are: (1) Policy-makers and practitioners lack expertise in how to interpret or understand the findings of research; (2) Policy-makers and practitioners lack expertise in how to apply the results of research to policy problems. Dissemination Importance of tailoring research when end-users are the focus This index is based on the importance attributed to various aspects of tailoring research when the focus is on end-users. This index is comprised of seven dimensions that range on a 6-point scale of adaption, ranging from 0 (does not apply), 1 (very 21 unimportant) to 5 (very important). The seven dimensions are: (1) readability and use of comprehension of my reports and research articles; (2) specific, operational nature of conclusions or recommendations; (3) provision of data that can be analyses by end-users; (4) sensitivity to end-users’ expectations; (5) presentation of reports (graphics, colour, packaging); (6) on-time presentation of research findings to end-users; (7) attention to ‘deliverables’. Importance of meetings & dissemination activities with end-users This index is based on the importance attributed to organising meetings and dissemination activities for end-users when carrying-out research. This index is comprised of four dimensions that range on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The four dimensions are: (1) preparing and conducting meetings in order to plan the subject and scope of projects with end users; (2) regular formal meetings to report on a study’s progress with end-users; (3) formal meetings to discuss findings with end-users; (4) preparing and implementing research dissemination activities for end-users. Importance of using presentations and reports to present research to parliamentary committees This index is based on the importance attributed to using methods such as presentations and reports for presenting research to parliamentary committees. This index is based on two items measured on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The two items are: (1) presentations to parliamentary committees and (2) sending reports to parliamentary committees. Importance of using the media to present research This index is based on the importance of using media to present research conducted by academics. This index is based on three items measured on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The three items are: (1) participation in radio and/or television programs; (2) publication of articles in non-academic outlets; (3) publication in electronic media, e.g. blogs and other social media. Importance of using refereed publications to present research The Journal variable is comprised of one item, publication of articles in refereed journals and is another method employed by academics to present or discuss their research. The importance of this method is measured on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). Interactions Importance of using contacts, seminar and This index is based on the importance attributed to using methods such as informal contacts, seminars and reports for presenting 22 reports to present research to policymakers and practitioners research to policy-makers and public practitioners. This index is based on six items measured on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The six items are: (1) informal contacts with policy personnel of government agencies; (2) informal contacts with public or community sector practitioners; (3) participation in seminars and workshops organised by government policy agencies; (4) participation in seminars and workshops organised by practitioners within public or community sectors; (5) sending reports to government policy agencies; (6) sending reports to practitioners within public or community sectors. Importance of using contacts, seminar and reports to present research to private sector organizations This index is based on the importance attributed to using methods such as informal contacts, seminars and reports for presenting research to private sector organizations. This index is based on three items measured on a 6-point scale, ranging from 0 (does not apply), 1 (very unimportant) to 5 (very important). The three items are: (1) informal contacts with personnel of private sector organisations (2) participation in seminars and workshops organised by private sector organisations (3) sending reports to private sector organisations. Number of external partnerships This index is the sum of all the partnerships with external organisations that academic researchers have engaged in. Role of the funding source: This project was finically supported through the Australian Research Council Linkage project LP100100380. This project has received cash and in-kind support from the following industry partners: Australian Productivity Commission; Australian Bureau of Statistics; Queensland Health; Queensland Dept of Communities; Queensland Dept of Employment; Queensland Dept of Premier and Cabinet; Victorian Dept of Planning and Community Development; Victorian Dept of Education & Early Childhood and the Victorian Dept of Human Services. The analysis and interpretation of the data presented in this paper was undertaken solely by the authors. The Industry Partners did not have any involvement in the study’s design or the survey data collected from academic researchers in Australian Universities. References Bell, M., Cordingley, P., Isham., C. & Davis., R. (2010) Report of Professional Practitioner Use of Research Review: Practitioner engagement in and/or with research. Coventry: CUREE, GTCE, LSIS & NTRP. Available at: http://www.curee--‐ paccts.com/node/2303. 23 Belkhodja, O., Amara, N., Landry, R., & Ouimet, M. (2007). The extent and organizational determinants of research utilization in Canadian health services organizations. Science Communication, 28(3), 377-417. Biesta, G. (2007). Why 'what works' won't work. Evidence-based practice and the democratic deficit of educational research. Educational Theory 57(1), 1-22. Bogenschneider, K., & Corbett, T. J. (2010). Evidence-Based Policy Making: Insights from Policy Minded Researchers and Research-Minded Policymakers. New York: Routledge. Burkhardt, H., & Schoenfeld, A. H. (2003). Improving educational research: Toward a more useful, more influential, and better funded enterprise. Educational Researcher, 32 (9), 3-14. Cherney, A., & McGee, T. R. (2011). Utilization of social science research: Results of a pilot study among Australian sociologists and criminologists. Journal of Sociology, 47(2), 144-162. Cherney, A. Head, B. Boreham, P. Povey, J. & Ferguson, M. (2011) The utilisation of social science research in policy development and program review. Preliminary report: Phase 1 Results. Institute of Social Science Research, University of Queensland Coburn, C., & Talbert, J. (2006). Conceptions of Evidence Use in School Districts: Mapping the Terrain. American Journal of Education, 112, 469-495. Cooper, A., Levin, B. and Campbell, C. (2009) The growing (but still limited) importance of evidence in education policy and practice. Journal of Educational Change, 10: 159-171. Cordingley, P. (2008). Research and evidence-informed practice: focusing on practice and practitioners. Cambridge Journal of Education, 38(1), 37-52. Cousins, J. B., & Leithwood, K. A. (1993). Enhancing knowledge utilisation as a strategy for school improvement. Knowledge: Creation, Diffusion, Utilisation, 14 (3), 305-333. Cousins, J. B., & Simon, M. (1996). The nature and impact of policy-induced partnerships between research and practice communities. Education Evaluation and Policy Analysis, 18(3), 199218. Davies, H. T. O., & Nutley, S. M. (2008). Learning more about how research-based knowledge gets used: Guidance in the development of new empirical research, working paper. New York: William T. Grant Foundation. Edwards, A., Sebba, J., & Rickinson, M. (2007). Working with users: some implications for educational research. British Educational Research Journal, 33 (5), 647-661. Fullan, M. (1992, June). Teachers as critical consumers of research. Washington, DC: Paper commissioned by OECD for Seminar on Producers and Consumers of Research. Gibbins, M., Limoges, C., Nowotny, H., Schwatsman, S., Scott, P., & Trow, M. (1994). The new production of knowledge. London: Sage. Head, B. W. (2008). Three lenses of evidence-based policy. Australian Journal of Public Administration, 67(1), 1-11. Hammersley, M. (2007) Introduction. In: M. Hammersley (Ed) Educational Research and Evidence‐ based Practice. London; SAGE. Hemsley-Brown, J., & Sharp, C. (2003). The use of research to improve professional practice: a systematic review of the literature. Oxford Review of Education, 29 (4), 449-471. Hess, F. M. (Ed.). (2008). When research matters: How scholarship influences education policy. Cambridge, MA: Harvard Education Press. Hillage, J., Pearson, R., Anderson, A., & Tamkin, P. (1998). Excellence in research on schools (Report no. 74). London: Department for Education and Employment. Holbrook, A., Ainley, J., Bourke, S., Owen, J., McKenzie, P., Misson, S., & Johnson, T. (2000). Mapping educational research and its impact on Australian schools In Department of Education Training & Youth Affairs (Ed.), The Impact of Educational Research, (pp. 15-278). Canberra: Commonwealth of Australia,. Honig, M. I., & Coburn, C. (2008). Evidence-based decision making in school district central offices toward a policy and research agenda. Educational Policy, 22(4), 578-608. 24 Huberman, M. (1990). Linkages between researchers and practitioners: A qualitative study. American Educational Research Journal, 27(2), 363-391. Kirst, W. M. (2000). Bridging education research and wducation policymaking Oxford Review of Education, 26 (3-4), 379-391. Knott, J., & Wildavsky, A. (1980). If dissemination is the solution, what is the problem? . Knowledge, Creation, Diffusion, Utilization, 1(4), 537-578. Landry, R., Amara, N., & Lamari, M. (2001a). Climbing the ladder of research utilization - Evidence from social science research. Science Communication, 22(4), 396-422. Landry, R., Amara, N., & Lamari, M. (2001b). Utilization of social science research knowledge in Canada. Research Policy, 30(2), 333-349. Landry, R., Lamari, M., & Amara, N. (2003). The extent and determinants of the utilization of university research in government agencies. Public Administration Review, 63(2), 192-205. Lester, J. P. (1993). The utilization of policy analysis by state agency officials. Knowledge: Creating, Diffusion, Utilization, 14(3), 267-290. Lester, J.P., Wilds, L.J., 1990. The utilization of public policy analysis: a conceptual framework. Evaluation and Program Planning 13, 313–319. Levin, B. (2011). Mobilising research knowledge in education. London Review of Education, 9 (1), 1526. Levin, B., & Edelstein, H. (2010). Research, policy and practice in education. Education Canada, 50(2), 29-30. Lingard, B. (2011). The impact of research on education policy in an era of evidence-based policy. Public Lecture, The University of New South Wales, 18 November, 2011. Lomas, J. (2000). Using linkage and exchange to move research into policy at a Canadian foundation. Health Affairs, 19, 236-240. Nutley, S., Walter, I., & Davies, H. (2007). Using evidence: How research can inform public services. Bristol: Policy Press. Oancea, A. (2005). Criticisms of educational research: key topics and levels of analysis. British Educational Research Journal, 31(2), 157-183. Orland, M. (2009). Separate orbits: The distinctive worlds of educational research and policymaking. In G. Sykes, B. Schneider, D. Plank, & T. Ford (Eds.), Handbook of education policy research (pp. 113-128). New York: Routledge. Ouimet, M., Landry, R., Ziam, S., & Bedard, P. (2009). The absorption of research knowledge by public civil servants. Evidence and Policy, 5(4), 331-350. Rickinson, M., Sebba, J., & Edwards, A. (2011). Improving research through user engagement. London: Routledge. Saha, L. J., Biddle, B. J., & Anderson, D. S. (1995). Attitudes towards educational research knowledge and policy-making among American and Australian school principals. International Journal of Educational Research, 23(2), 113–126. Vanderlinde, R., & van Braak, J. (2010). The gap between educational research and practice: Views of teachers, school leaders, intermediaries and researchers. British Educational Research Journal, 36(2), 299-316. Weiss, C. H., & Bucuvalas, M. (1980). Social Science Research and Decision-Making. New York: Columbia University Press. Wilkins, R. (1988). Research and policy in teachers’ organizations. Journal of Education Policy, 3(2), 89-103. Zeulie, J. (1994). How do teachers understand research when they read it? Teaching and Teacher Education, 10(1), 39–55. 25 Table 1. Research Utilisation Scale Variable Transmission Cognition Reference Effort Influence Application I transmit my research results to end-users My research reports have been read and understood by end-users My work has been cited in reports and strategies by end-users Efforts were made to adopt the results of my research by end-users My research results have influenced the choices and decisions of end-users My research has been applied by end-users Table 2. Proportion of education respondents at each stage of the research utilisation scale n % No echelon 6 3.9 Transmission 3 1.9 Cognition 16 10.3 Reference 11 7.1 Effort 2 1.3 Influence 1 0.6 Application 117 75.0 156 100.0 26 Table 3. Internal reliability coefficients (Cronbach’s alpha) for variables – Education research Name of variable RU Index Researchers’ Context Research targeted to user Importance of academic funding Importance of other funding Benefits of collaborative research Barriers academics experience in the transfer & uptake of their research Consequences of investing in research partnerships Problems relating to the orientation of research partnerships User’s Context End-users prioritise high quality research End-users prioritise the useability of the research End-users prioritise the feasibility of the research Policy-makers have found my policy focussed research to be relevant Barriers policy-makers experience in transfer & uptake of research Dissemination Importance of tailoring research when end-users are the focus Importance of meetings & dissemination activities with end-users Importance of using presentations and reports to present research to parliamentary committees Importance of using the media to present research Interactions Importance of using contacts, seminar and reports to present research to policy-makers and practitioners Importance of using contacts, seminar and reports to present research to private sector organizations Number of cases Number Cronbach of items alpha in a scale 156 6 0.77 156 156 156 156 156 4 2 5 10 5 0.60 0.34 0.78 0.85 0.74 156 156 10 5 0.85 0.81 156 156 156 156 7 4 3 3 0.72 0.72 0.68 0.97 156 2 0.95 156 7 0.69 156 4 0.88 156 2 0.89 156 3 0.62 156 6 0.79 156 3 0.87 27 Table 4. Means and standard deviationsa Education research Range Research Utilization Ladder Research Approach Quantitative studies Qualitative studies Researchers’ Context Research targeted to user Importance of academic funding Importance of other funding Benefits of collaborative research Barriers academics experience in transfer & uptake of research Consequences of investing in research partnerships Problems relating to the orientation of research partnerships Research Time (teaching and research or research only positions) Number of external partnerships User’s Context End-users prioritise high quality research End-users prioritise the useability of the research End-users prioritise the feasibility of the research Policy-makers have found my policy focussed research to be relevant Barriers policy-makers experience in transfer & uptake of research Dissemination Importance of tailoring research when end-users are the focus Importance of meetings & dissemination activities with endusers Importance of using presentations and reports to present research to parliamentary committees Importance of using the media to present research Importance of using refereed publications to present research Interactions Importance of using contacts, seminar and reports to present research to policy-makers and practitioners Importance of using contacts, seminar and reports to present research to private sector organizations Number of external grants a. Min Max Mean 1 5 3.64 Std. Err. 0.05 0 0 1 1 0.40 0.83 0.04 0.03 1 0 0 0 0 0 0 4 5 5 5 5 5 5 2.37 4.03 3.49 3.36 3.90 3.68 2.67 0.04 0.06 0.07 0.07 0.06 0.06 0.07 0 1 0.81 0 35 6.22 0.49 1 1 1 5 5 5 3.91 4.64 3.90 0.05 0.03 0.06 0 5 3.54 0.12 0 5 3.23 0.08 0 5 4.24 0.04 0 5 4.21 0.05 0 5 2.41 0.12 0 0 5 5 3.20 4.71 0.07 0.05 0 5 3.77 0.06 0 5 2.91 0.10 0 44 7.76 0.56 Standard deviations only reported for continuous measures. 28 Table 5. Regression equations predicting utilisation of education research M Research Approach Quantitative studies 0.14 Qualitative studies -0.11 Researchers’ Context Research targeted to user -0.02 Importance of academic funding -0.06 Importance of other funding -0.05 Benefits of collaborative research 0.19*** Barriers academics experience in transfer & uptake of research -0.06 Consequences of investing in research partnerships -0.20** Problems relating to the orientation of research partnerships 0.15* Teaching & research position 0.05 Number of external partnerships -0.01 User’s Context End-users prioritise high quality research -0.05 End-users prioritise the useability of the research 0.25* End-users prioritise the feasibility of the research -0.14* Policy-makers have found my policy focussed research to be 0.07* relevant Barriers policy-makers experience in transfer & uptake of research 0.03 Dissemination Importance of tailoring research when end-users are the focus 0.34** Importance of meetings & dissemination activities with end-users 0.03 Importance of using presentations and reports to present research to -0.00 parliamentary committees Importance of using the media to present research -0.01 Importance of using refereed publications to present research -0.06 Interactions Importance of using contacts, seminar and reports to present 0.13 research to policy-makers and practitioners Importance of using contacts, seminar and reports to present 0.02 research to private sector organisations Number of external grants 0.02*** Constant 1.28 Observations 155 Adjusted R2 0.28 SD (0.10) (0.14) (0.10) (0.06) (0.06) (0.07) (0.08) (0.10) (0.09) (0.13) (0.01) (0.10) (0.13) (0.07) (0.04) (0.05) (0.14) (0.11) (0.04) (0.06) (0.08) (0.08) (0.04) (0.01) (0.79) Standard errors in parentheses * p < 0.10, ** p < 0.05, *** p < 0.01 29 Figure 1. Number of education researchers climbing the echelons of the ladder of knowledge utilisation – progression is subject to passing previous echelons 160 147 150 Number of researchers 140 131 120 118 120 117 100 80 Pass Fail 60 40 20 16 6 11 3 2 0 Transmission Cognition Reference Effort Influence Echelons of the ladder of knowledge utilisation 1 Application Figure 2. Number of education researchers passing each stage of research utilisation – failure in one stage does not preclude passing subsequent stages 160 148 150 140 135 133 128 134 Number of researchers 120 100 80 Pass Fail 60 40 28 21 20 6 23 22 8 0 Transmission Cognition Reference Effort Influence Stages of research utilisation Application 30