Ref 8.10 The evaluation of training and HRD: A critical study with applications. Referred paper Eduardo Tomé, Universidade Lusíada de Famalicão Email: eduardo.tome@clix.pt Keywords: Training, Evaluation, HRD, Economics. Abstract: In this paper we review the evaluation methods used on training and HRD programs. Specifically we address two questions in succession. First we try to assess if some methods are better than others on defining the true impact of HRD and training investments in the economy. Secondly we use a very specific methodology (Gramlich 1990) to evaluate some very well known experiences on the evaluation of training. In that methodology, the benefit derived from the operations depends on three factors: 1) the probability that the methodology defines correctly the program impact; 2) the probability that the administrator takes on account the evaluators’ conclusions; 3) the total value of program expenses. 1 Introduction Societal background: In the 21st century economy it is well understood that Human Resources are a major factor of economic and social wellbeing. In world terms, the Education Index, computed by the United Nations Development Program (UNDP) shows the clear relation between Education, income and Human development (Table 1). The Human Development Report (HDR) Education Index is obtained computing the adult literacy rate (% for the population aged 15 years old and above) and the combined gross enrolment ratio for primary, secondary and tertiary education in percentage. Table 1 - Education, Income and Human Development – 2006 Countries Income Human Development 33082 0.936 High Income 7416 0.776 Middle Income 2631 0.570 Low Income 9543 0.743 World Source: United Nations, 2008 Education Index 0.937 0.843 0.589 0.750 We must note, that the HDR Education Index is not only a measure of the level of Education in countries, but it is also a result of decades of investment in Education and in basic HRD in those countries. In what concerns specifically training, it is well known that the more developed countries are also those that invest more in private training (funded by companies and by individuals) and in public training (funded by the state). Usually those investments are measured by the share of adults that participate in life long learning, and by the share of individuals that undergo training, or by the funds spent in active labour market policies (see Table 2). Table 2 - Investment in Training: Basic Statistics Life Long Learning (2006) Public Euros 7.5 608 Belgium 1.3 14 Bulgaria 13 Czech Republic 5.6 29.2 1059 Denmark 7.5 5580 Germany 6.5 4 Estonia 7.3 342 Ireland 1.9 71 Greece 10.4 1334 Spain 7.6 4927 France 6.1 2839 Italy 6.9 13 Latvia 4.9 11 Lithuania 3.8 34 Hungary 2 Funds (2005) Million 15.6 723 Netherlands 13.1 800 Austria 4.7 250 Poland 4.2 431 Portugal 1.3 10 Romania 15.0 12 Slovenia 4.1 9 Slovakia 23.1 581 Finland 32.0 981 Sweden United 26.6 1559 Kingdom 18.7 891 Norway Source: Eurostat 2007 Note: Life Long Learning refers to the percentage of the population aged 25 to 64 participating in education and training over the four weeks prior to the survey. Public funds refers to the Public expenditure on training for unemployed and groups at risk The data related training are usually collected regarding the OECD (OECD, 2006), the EU (Eurostat, 2007) and the other developed countries, with few exceptions (UNESCO 2006). However, even if the investment in HRD and training is seen as an obvious necessity, the evaluation of that investment poses many crucial questions. First of all resources are scarce and must be used with the best return possible. Secondly, in order to guide policies, it is essential to evaluate investments. That evaluation might be made in macroeconomic terms, if the program has a macroeconomic impact, but must be made at micro level in any case. Therefore, it is crucial to know which the best methods available to evaluate training and HRD are. The evaluation problem has a particular significance, because with the emergence of a Knowledge Based Economy (KBE) the societal needs for training and HD worldwide are not decreasing, quite on the contrary, they are increasing. Finally, every person who had some important position in the administration of HRD and training programs will tell you that the evaluation processes put in place, are dependent on the assumptions and feelings the administrators have on the programs. Putting it very basically, we only evaluate when we have doubts. We don’t evaluate if we are sure that the program is very good (or very bad). Research Questions: Given what was written above, this paper will address two basic research questions: 1) What are the best methods available to define the impact of training / HRD in the economy? 2) How can we evaluate the evaluations that have been made, having in consideration the dimension of the program, the evaluation method that is used, and the aprioristic conceptions of the administrators ? Plan of the paper: In the first part of the paper we define the theoretical background. Namely, we review the basic economic theories on training and HRD (1.1) and we introduce the evaluation problem (1.2) and describe the most important evaluation methods on the topic (1.3), dividing them in result methods (1.3.1), and impact methods (1.3.2). In the second part of the paper we present a model (Gramlich, 1990, 160-4) that we consider is 3 fundamental to answer to the two questions we posed. In the third part we describe data we collected on evaluation experiences from some of the most important HRD and training programs all over the world. In the fourth part, first (4.1) we analyse the methods described in 1.2 according to the probability of defining the true impact of the program. Therefore we answer to question 1. Secondly, in section (4.2) we apply the Gramlich’s method to the experiences we described in section 3. In section 5 we discuss the paper findings (5.1), point the papers’ limitations (5.2) and suggest further studies (5.3). Finally, in section 6 we present the paper’s conclusions. Theoretical background Theories The positive influence of training and HRD in the situation of individuals, companies, and nations was noticed by the most important thinkers in the economic world, since the dawn of Economics. Indeed, Adam Smith, was the first to notice that skills and competences are an important factor in the development of countries (Paul, 1989, 59-60). The other most important classical economists agreed with Smith’s idea: Marshall (Paul, 1989, 59-60), Stuart Mill (Shackleton 1995, 10), Pigou (Stevens, 1996, 21). Keynes, also gave competences and skills an important place in the economic process (Keynes, 1990, 191). But it is very well known that the basic economic theory that is nowadays considered on HRD, and E & T saw the light with the works of neoclassical economists in the 50s, 60s and 70s: Becker, 1962; Becker 1976; Becker, 1980; Ben-Porath, 1967; Mincer 1958; Mincer 1962; Mincer 1974; Oi, 1962 and Schultz, 1971. According to those studies, HRD should improve labour productivity, and in consequence the wages of qualified people should be higher than those of non qualified people. Furthermore, HR being a form of investment, its returns should essentially be perceived in the long run. And, given that according to Human Capital Theory (HCT) the process of economic development was to be based in HR, at least in the long run a higher level of HR was considered to be a condition for employment. Those analysis were quite strong and valid, but also a bit simplistic. And it is quite interesting to note that in the last decades, HCT theories have been challenged in significant a number of ways, and that, as a consequence of the critical contributions that challenged the HCT, the corpus of economic theories on HRD grew in stature and the real world was much better explained. Some of the major advances relate to the explanation of how discriminative factors like race, gender and company dimension may be incorporated in the analysis, to the public intervention in the labour market, and to the role of knowledge in the economy. We will describe succinctly these contributions in the next paragraph. Regarding the discriminative factors, critical analysts explained why some classes (among them ethnic minorities, adult women,, disabled persons) are given less HRD opportunities because they are perceived beforehand (screened) by companies as being less productive; and on the contrary people with diplomas and “politically correct” status are given all the training opportunities by organizations because they are perceived as productive. As a consequence, inefficiencies and inequalities are originated (Spence, 1973; Stiglitz, 1975). This analysis originated a vast bibliography on contracts and asymmetric information in the labour market, to deal with that market imperfection. Other important fact is that not all the companies that act in the labour market are equal. Quite on the contrary there is a tendency for the market to be quite dualistic, with small 4 SMEs having difficulties training employees, and big companies offering solid career perspectives and big training opportunities. As a consequence training tends to be better remunerated in big companies than in small ones (Chapman, 1994; Doeringer and Piore, 1985). Regarding the public intervention in the labour market, for HCT theorists, HR investments were mainly private investments, made by workers and companies; but, quite a number of studies have since suggested that the public intervention in the HR arena should be made, at least to eliminate some market failures related to externalities, market imperfections, and weak private capacity to train (Middleton, Ziderman and Adams, 1991, 35) funding difficulties (Katz and Ziderman, (1990), transferable training (Stevens, 1996), the bad job, low skills trap (Snower, 1996), and individual myopia (Dougherty and Tan, 1997, 38). Those studies also pointed out that in case of equity problems the public intervention could be justified (Booth and Snower, 1996). Of course, as a general golden rule, the public intervention in the HRD market should only be made if the dimension of the government failure it carries with, is smaller than the dimension of the public failure that is to be solved (Booth and Snower, 1996). It is also very curious to note, that several philosophic schools exist in social policy, and we can assume that they differ in the relative dimension they give to market failures versus government failures (see Table 3) (Alcock, Erskine and May, 2003; Esping Andersen, 1990; Ferrera and al, 2000; Deacon, 2000). An application of those differences of conception to the Human Resource Management reality was already done (Ignatovic and Svetlik, 2003). Finally, a vast number of studies have been done on Knowledge Management, that in some way extend the analysis of the use of HR in companies, following the studies of Nonaka and Takeuchi (1995), and Kaplan and Norton (1996). Table 3: Public Intervention versus Private Intervention in HRD Type of Social Market Government Consequence System Failure Failure for HRD Minimal Maximal Privatized Liberal Moderate Considerable Subsidiary Conservative Considerable Moderate State led Social Democratic Maximal Minimal Nationalization Socialist Example UK, USA Central Europe Nordic Countries Eastern Countries Source: Own analysis of Alcock, Erskine and May, 1997; Esping Andersen, 1990; Ferrera and al, 2000; Deacon, 2000 The problematic about evaluation Program evaluation is a complex question. At least 10 questions can be put when evaluating a program (see Table 4). Table 4: Questions that might be addressed when evaluating a program Question How ? When ? Where ? Meaning Methodology Timing Sampling, Target Group, Organization, 5 Country Rationale Planner, Evaluator Objectives Decision Maker Decisions Criteria Beneficiary Why ? Who ? What ? Whom ? What decisions ? Which criteria ? Which benefit ? Source: Chinapaj and Miron 32, 1990. In this paper we concentrate ourselves in the first and in the last question, namely the methodology and the benefit from the evaluation. However, some other questions will also be important in the discussion, namely number 3 (Where, When). The second question (Why) was analysed in the first section (Theories). Evaluation Methods Economists and managers have developed many important evaluation methods in order to evaluate training and HRD. Those may be analysed in the context of the evaluation of social policies (Tomé 2005). Accordingly, some of those methods relate basically to occurrences (they may be called results methods), and some of those methods relate to the perceived consequences of the training programs (they may be called impact methods). We will present result methods (1.3.1) and impact methods (1.3.2) in order to analyse the methods used in evaluations (see 4). Result Methods The easiest way to evaluate HRD investments or policies is to count the level of investment in money terms (financial indicator) and the number of people that participate in the operation, namely the trainees (physical indicator). However result indicators only analyse costs and participants, and are not related with benefits. At least in theory it is absolutely possible that a large program (regarding costs and participants) may have very low or even negative consequences for the people involved. But, result methods tend to be used to evaluate policies because they are cheaper, less risky, less random and much easier to command than impact methods (Tomé 2005). Impact Methods These type of methods try to define the difference the program makes to the society in which it is implemented (Tomé 2005). There are several ways of putting that idea in practice. In economics and social policy, these methods involve computing the difference between the evolution of a group of participants and a group of non participants. In a weaker form, some studies only analyse the difference “before-after” regarding the group of participants. When two groups are compared, selection problems may exist. Therefore two main sub-groups of impact methods exist: experimental studies in which the selection to the program is done randomly, and quasi-experimental studies in which the selection to the program is not done randomly but a selection rule is estimated to eliminate the selection bias. If we consider that there is not a selection bias (as in experimental studies) we use simple control function estimators. If we consider that the selection bias is only originated by the observable variables (as in non experimental 6 studies), we may use first difference estimators. If we consider that the selection bias is originated by non observable variables (also in non experimental studies) we have to use difference on difference estimators (Heckman and Hotz, 1989). Impact methods are very important because they are the finest way to detect the program benefits, and therefore to make cost benefits analysis. But they present quite a number of difficulties: they are expensive (in money, time, resources, decisions), they are risky (because they may expose the programs’ and the administrators failures), and they are random (because there is no agreement within the scientific community about what are the best estimators, and in the past different methods have originated different impact estimators) (Tomé, 2005). HR scholars have tried to measure the impact of a program by interview and questionnaire methods, above all by the Kirkpatrick model of four levels (Kirkpatrick, 1998). The four levels are immediate reaction, learning, behaviour and organizational results. Traditional managers and accountants try to define the impact by converting HRD in a Return of Investment (ROI) perspective (Fits-zen, 2000). And in the last decade knowledge managers have evaluated HR investment in the perspective of Intellectual Capital investment using the Balance Scorecard methodology (Kaplan and Norton, 1994). Regarding all those methodologies (Schellhass, 1991) it is important to stress that from there is a difference between the optimal impact of the program (A) which tend to be less then the real effects (B). However, for political and administrative purposes the predicted effects (C) may be estimated be default originating that in the end the real effects tend to be higher than the predicted ones (C > B). However, the evolution of the control group or comparison (D)has to be taken into account, and the impact of the program (E) is given by B – D. Different perspectives of the evaluation It is quite important to verify that not all the methods that are used evaluating training and HRD programs have the same perspective. In our opinion at least five different perspectives exist, and every one of them results in different evaluation methods, namely the following: 1) the public administrator perspective; 2) the private manager perspective; 3) the HR expert perspective; 4) the accountant perspective; 5) the economist perspective: 1) In the public administrators’ perspective the best way of evaluating a program is to use basic result methods. These methods provide the administrator with the description of the operation. If the operation was received by a large number of people and organizations, this is meant to be a success. According to this way of thinking the final report should include financial and physical results. The public administrator may also sense the importance of assessing the employability of the program. Other two very important notions are the difference between the expected evolution and the real evolution, and the ex-post assessment that the program really met social needs. That is, public administrators are concerned with the execution of the program and the correctness of the planning. The definition of the impact of the program is only a second thought for the administrators; 2) In the private managers’ perspective HRD and training programs are looked as ways of spending own money; so, the evaluation of that investment might be made almost intuitively by the evolution of the company’s overall performance. However, sometimes the manager may want to test strictly the “value for money” aspect of the 7 investment in HR and in these circumstances ROI evaluation methods are the most preferred. The basic handicap with those methods is that they overlook all the evolution outside the company, and a similar value of profitability may mean very different things in times of crisis or in times of growth; 3) In the HRD expert perspective, the best ways of evaluating the programs, are those that deal more immediately with the training scene; there fore HRD managers tend to prefer Kirkpatrick reaction models or competence based models of evaluations in which questionnaires or interviews are applied or even the methods in which the observer participates closely in the operation; these models have also been developed in the knowledge management analysis; A very significant branch of science was developed analysing knowledge creation, knowledge sharing, knowledge transfer, and knowledge audit; 4) The accountancy perspective began to be very important in the last two decades, when it became obvious that the company value was not dependent only and essentially on the tangibles, but was mostly dependent on the value of intangibles; the Balance Scorecard methodology was made in order to solve the equation: Market Value – Book Value. Another branch of analysis was developed around the concept of Intellectual Capital; 5) Finally, in the economist perspective, in particular in the HRD economists perspective, the most important is to define the impact of private or public policies, and microeconomic impact models, or macroeconomic input-output models are the way to detect the program impacts. Table 5: A typology of evaluation methods Perspective Method Classification Participants, Costs, Results Public administrator Direct Employability ROI Impacts for company Private manager Interviews, Impacts for agents HRD expert involved Questionaire, Participant – Observer Kirkpatrick, Finances, Structural Impacts for company Accountant Capital, Social (finance, structures, Capital, Human costumers, human Capital capital) Control Group Impacts for society Economist Input Output Methods Branch of science Public administration Private Finance Knowledge Management Intellectual Capital Macroeconomics, Human Resource Economics How to evaluate an evaluation We think that nowadays evaluation is a common procedure in developed countries and in conscious and prosperous institutions. However, we also think that not all the efforts made to evaluate training and HRD have the same quality, and therefore, the evaluation of the evaluation is a very important scientific topic. A very interesting model to evaluate evaluations was presented some years ago (Gramlich, 1990, 160-4). In this model the benefit derived from an evaluation is defined as a function of three variables: 8 BEN = p * q * PV The meaning of each one of those variables is the following: 1) p is the probability of discovering the true impact of the program; 2) q is the probability that the government / private authority will pay attention to the result of the evaluation and take it into account. 3) PV is the program value measured in monetary units. Using this method of evaluating evaluations we arrive to the following conclusions concerning the priorities regarding program evaluations : 1) it is better to evaluate using very fine, even if extremely costly impact methods (for which p is large); 2) it is better to evaluate large programs (for which PV is large); 3) it is better to evaluate programs when there is a doubt about the program usefulness (and therefore q is large). It is also interesting to understand that, in a complementary perspective (Burtless and Orr, 1986) the benefit of an evaluation can be measured as the costs (F) that are saved if the program is stopped minus the benefits that cease to exist if the program is stopped (G) plus the evaluation costs (H). It is clear that if F is higher than G + H the evaluation is socially beneficial. Therefore, high values of p, q and PV may generate a high F, that will ensure that G plus H are offset. In this circumstance the evaluation is worth. However, a costly evaluation (with an high H) and a high F, may identify an even higher G and in this hypothesis, H is justified because it showed that G exists and G is an important and positive social benefit. Known experiences of evaluation analysed and evaluated We will separate the evaluations on training geographically. In section 3.1 we will describe the American cases, in section 3.2 we will deal with the European cases, and in section 3.3 we will address cases from other continents. The variables that we will use to study PV, q and p are basically described in the following Table 6: Table 6: Elements of the model and relevant proxys Concept Relevant proxys Physical indicators PV Financial indicators Political discourse Q Type of methodology put in place Detailed analysis of the methodologies P used The American case Public sponsored training and HRD programs Very important evaluations on public sponsored training programs have been made in the USA since the sixties (Tomé, 2001). Initially those evaluations were above all made using result studies. But since the final years of the seventies impact studies began to be performed on the those programs (Ashenfelter, 1978). The programs that 9 were evaluated include those made under the Manpower Development Training Act (MDTA) and the three Acts that followed that piece of legislation in succession: the Comprehensive Employment Training Act (CETA), the Job Training Partnership Act (JPTA) and the Work Investment Act (WIA). A detailed list of the evaluations made to the training programs in the United States is included in (Lalonde 1996). It is quite interesting to study the evolution of the evaluation of those programs, in special those four major programs, because in the mid eighties the CETA program was subjected to a theoretical discussion in the Academy of Sciences of the United States (Ashenfelter and Lalonde 1996). The most important part of the discussion was related with the fact that very different methods of non-experimental evaluation had originated very different estimates of the program impact. That troubling finding had a crucial consequence: the JTPA program was evaluated with an experimental study (Orr, 1996). In consequence we may say with a very high degree of certainty that in relation with the American case of public sponsored training and HRD programs, p was large, q was large, PV was also large and the benefit from that evaluation was also very big. Privately sponsored training and HRD programs The evaluation of HRD in the private programs is a common occurrence. Private training and HRD programs have been made in the USA by the most influential companies and organizations since WWII. Evaluation of HR investment has been also made by many smaller companies whenever they want to assess HR investments. HR and training have been evaluated using a plethora on methods, namely the following (Mc Lean 2005, Smith and Piper, 1990): questionnaires: testing and griding; Kirkpatrick four levels; trained observers: interviews; return on investment methods; impact microeconomic methods; Balance Scorecard; systemic approaches. It is very curious to note that training was widely evaluated, and that the values of the programs in question were very high. Those programs were evaluated using pedagogic methods, traditional management methods and new accountancy methods. So, in relation to Gramlich’s methodology, PV, b, q and BEN are considered to by high, even if impact studies were seldom used, because the evaluation approach was company based and not program / society based. The European Case The programs funded by the European Union The EU began funding training and HRD in 1960. However, until 1989 the programs funded by the European Social Fund were only evaluated by result indicators. This happened because from the perspective of the European Union (at the time known as European Community) the simple fact of awarding money to the less developed regions of the EU, and to some target groups was seen as beneficial to those regions and persons. Therefore physical and financial indicators were seen as good indexes of the benefits derived from the programs. Until 1972, each country new the difference between the money received and the money spent paying the ESF operations, and the operations were only funded ex-post, meaning that there was a direct link between the money spent and the level of employability: only successful operations were funded. From 1973 to 1989, with the establishment of ex-ante payments, the awarding of funds was money according to some “Orientation Guidelines”. In consequence when 10 the demanded amounts exceeded the available amounts, the proposed value of funds was reduced in percentage even if they were eligible. The evaluation of the operations was made by accounting the number of persons supported and the money spent. No impact studies were performed. The ESF also asked for the situation of the trainees six months after the operations but those data were not made public. Calculations were possible on the relation between the amount of money awarded in the candidature phase and the amount of money. It was also possible to analyse the number of trainees per 1000 habitants at the EU level and the number of ECUs per inhabitant at EU level also. We call that way of analysing the “absorption analysis”. In consequence we must say that before 1989, regarding the EU, PV was increasing, put p was small, and q was even smaller. The potential benefit derived from that type of evaluation was very small. Table 7: Evaluation experiences according to the Gramlich model PV P Q Country Funding Programs Public MDTA, High: High: Very High: USA CETA. Billions of Impact Concern CETA, dollars studies over WIA evaluation reaches Academy of Sciences Private Main private Very High : High: High: companies Many Plethora of Concern of Billions of scientific and with methods but economic organizations. dollars Small few impact impact was companies studies. evident. when Concern of concerned impact as with HR calculated development. by Kirkpatrick, ROI and Balance Scorecard was evident too. However, few impact studies. Until 1990 High and Small – Small – Europe Public EU increasing – results and Spending budget absorption and augmenting absorption are indicators of benefits. From 1990 to High and Rare – very Medium – real Legal 2000 increasing – few budget evaluation concern 11 BEN Very high: billions of dollars invested and assessed. Very High: Huge investment assessed on daily basis. Very small – the evaluation was not valid. Very small – in practice almost augmenting Other studies performed Increasing – impact studies made on the European Strategy for Employment High: impact studies in growing number of cases Since 2000 Very high and still augmenting Public Sweden, France, Spain, UK Very high and increasing through time Private Major private Very high: High – companies billions Economic and since WWII and organizations. economic Some small methods companies. even if not using impact studies Eastern High – Small – countries Training Existing as a was a form of widespread positive phenomenon evaluation Public with evaluation High – Legal concern with evaluation. Evaluation reports High – need to assess scientifically the use of public money and its economic impact High – Always evaluated, even if with company based concerns, case studies Very small – Taking for granted that to the program was positive nothing changes Medium – countries that use impact studies are not the majority High – Investment scientifically assessed. High – Investment frequently assessed. Very small – Almost impossible to detect any impact Public Brazil, Canada, Australia High High – High – High socioconcern policy economic with public guidance studies money World Bank and NGOs Less Developed Countries Important Small – Small – Small – result beneficial international indicators instincts aid concerns dominate prevail Private Brazil, Canada, Australia, Japan High High – High – High – company concerns guidance of based about private methods usefulness investment 12 – In 1989, the European Commission tried to reform the Structural Funds (SFs). As a consequence a new procedure of implementation was defined. The SFs began to be organized in programs. For each program four operational phases were defined: ex-ante appreciation, monitoring, control, and ex-post evaluation. In consequence, the importance of the evaluation of programs was reinforced, at least in theory. And we may say that from 1990 to 1999, at the epoch of the first two “programming” phases from the Structural Funds reform (phase I 1990-3; phase II 1994-9), the value of PV was increased because the funds available for the ESF were augmented, and the value of q was also increased because the concern with the evaluation of programs was clearly shown in the legislation. However, in practice the desire of evaluation was expressed in evaluation reports that had the following structure: 1) Description of the situation before the program and justification for the program; 2) Description of the program and description of the differences between the expectations and the reality; 3) Inquiry in the people (trainees and trainers) and in the organization of evolution of the situation “before-after”. Therefore, the value of p was not considerably augmented, even if it was expressed that the effect of the programs was going to be detected. It is a clear sign of the situation that existed in the EU regarding the evaluation of training and HRD that in 2001, the European published the collection “MEANS”, named “Measures to Evaluate the Effectiveness of Structural Funds” (EU, 1999). That book was composed of five volumes and in them the European Commission made a summary of the various types of methods that were available to study the Structural Funds. As a consequence, it was expectable that those methods would be extensively put in use in the third programming phase (2001-6). And indeed, for some countries (namely Spain, Greece, France, Netherlands, Austria, Sweden) the investment in training by the EU was evaluated in 2002 using impact studies, as a part of the European Strategy for employment, and in some cases (Spain, Greece and Austria) the impact was found to be positive (EU 2002). One very interesting characteristic of the evaluations made by the ESF is that they are made ex-post on the whole program and not only ex-post on one of the years of the program. This characteristic in turn implies that it is almost impossible to perform a meaningful impact study because impact studies should be made when the program is still operating. So, as a general conclusion, we must say that in the European Union funded projects, the value of PV has been increasing steadily since 1958, and that increase was a consequence of the increasing importance HR, HRD and training had in the European economy. However, the value of p has not increased very much, in all the countries, because we do not think that the methodologies of evaluation that were used find the true impact of the program in socio-economic terms. Furthermore, we also think that regarding the other aspects of the evaluation (namely the pedagogical and knowledge based ones) p was low. Finally, we think that even if the concerns of the administrators rose with the time, the fact the methodologies used in the evaluation are basically basic result ones, that the evaluations compared the situation of the participants in the program in a before-after comparison, and that the retroactivity of the evaluation was severely diminished, means that q is growing and is high in the discourse but low in practice. The exception to this rule is the evaluation made in the scope of the European Strategy for Employment (EU, 2002). As a rule, BEN is small. 13 Programs funded by the National Budgets Those type of programs were subjected to a large set of different types of evaluation: 1) Basic result methods were widely applied in the generality of the programs put in place; 2) Methods related to the pedagogical aspects of HRD and training were also used in many cases; 3) Impact methods were used in some cases, in different countries, most of all in Sweden, Norway, France, the UK, Netherlands, Spain. The concern with the evaluation of training in those countries and regarding public funded programs was clear from the since the eighties, and since than more sophisticated evaluation methods were implemented. It is also clear that the value of the PV variable has been increasing with the passage of the time, because the economic changes meant that the public presence in the labour market, regarding training and HRD was widened and reinforced. So, we consider that in this case the values for the three variables that constitute the model are high, and BEN is also high. The programs funded by the private organizations Training and HRD practices have been a frequent practice in Europe since WWII and mainly since the sixties. Most of the privately funded programs have been submitted to simple forms of evaluations. Many programs have been evaluated on pedagogical and reactive grounds. Some programs have been evaluated with ROI techniques and with Balance Scorecard techniques. A small number of programs have been evaluated with impact studies. It seems that even if the organizations were quite concerned about training and HRD, a fact that is evident by the increasing number of workers involved in those types of operations, by the increase in the money spent, and by the large number of evaluations made. However, it is also quite interesting to see that most of the evaluations made were related to the company own results. It seems that the European private companies rely on the fact the HRD and training are useful, and that they expect that usefulness to be shown in terms of competences, financial results, routines or costumer satisfaction. So in this case, PV, p and q look to be really big even if impact studies of social scope are not used. The rare use of impact studies is justified by the fact that companies have a more inner oriented approach and that approach is more easily described by Kirkpatrick, ROI or Balance Scorecard methods. Other cases Public funded cases Important HRD and training programs funded by the public administration took place outside the USA and Europe in countries like Brazil, Australia, and the Eastern Countries of Europe. In what concerns the cases of the Eastern countries little statistical information is known, with the exception that in those countries training was a part of the educational system, and that in those countries the level of training and education was high, when compared to the world average. We also don’t know of any experience of evaluation in those countries, and we are led to the conclusion that training was made 14 by the public organizations because it was meant to be beneficial for the country, and that fact was enough to justify the occurrence of training. The only evaluation method was the existence of training itself. So, PV was large, p was very small and q was even smaller. The BEN derived from that type of evaluation was very small and that situation led to massive changes in the administration of training when the socialist regimes were replaced by democracies. In Australia, Brazil and Canada, large training and HRD programs funded by the public authorities existed since the sixties. Those programs were submitted to impact evaluations in some specific cases. In those cases it was quite clear that the program administrator was clearly concern and doubtful about the program impacts and wanted to check them. So, p was high, q was high, PV was high, and BEN was also high. World Bank and NGO funded cases In some LDC countries, the HRD and training programs have been done with the support of the World Bank and the other NGOs. Those programs were usually evaluated using result indicators, and competence based indicators. We don’t know of any deep and fine economic analysis regarding the use of those programs. We consider that in these cases PV was small, and p and q were also small, even if there was some justification in the fact that the promoters thought they were providing much needed social support to those societies. It is also possible that in the future in the context of the Millennium Goals those programs will be increased. Privately funded programs Privately funded training and HRD was also a common characteristic of developed countries, like Japan, Canada, Australia, and New Zeeland. As it happened in the USA and in Europe, those programs had an important dimension, and were frequently evaluated using complex scientific techniques. That fact happened because the companies were concerned with the return of the investment they made. Concluding comments Conclusions The use of Gramlich’s methodology to evaluate training methodologies and their practical use led use to three main conclusions: 1) Some very basic and administrative methods are clearly less effective detecting the true impact of the program; 2) However a very wide of scientific methodologies are available with the intention of detecting the “true” effect the program has; and we found that rather surprisingly all those methodologies are similar, because they try to answer the basic question on the “program” difference / consequence, put in different perspectives; 3) When we analysed the practical use of evaluations on HRD and training programs we found that State funded programs are evaluated with “socio-economic” impacts studies and privately funded programs are evaluated with indicators related to the company’s situation. The exception are EU programs in which the reliance on “European money” and “absorption” concepts decreases the level of impact analysis. 15 References Alcock, P.; Erskine A. and May M. (2003) The Student’s Companion to Social Policy. Second Edition. Blackwell Publishing. Ashenfelter, O. (1978), "Estimating the effect of training programs on earnings", Review of Economics and Statistics, Vol. LX pp.47-57 Ashenfelter, O., LaLonde, R. (Eds),The Economics of Training, Volume I, Theory and Measurement, Edward Elgar Publishing, Aldershot, pp.ix-xvi. Becker G. (1962) Investment in Human Capital: A Theoretical Analysis - The Journal of Political Economy – volume LXX - 5 - part 2 - Supplement - October – pp.9-49. Becker, G. (1976) The Economic Approach to Human Behaviour - The University of Chicago Press - Chicago. Becker , G. (1980) Investment in Human Capital: Effects on Earnings - in Human Capital : A theoretical and empirical analysis whit special reference to education Chicago - University of Chicago Press - Chapter II - pp.15-44. Ben Porath, Y. (1967) The production of human capital and the life cycle of earnings Journal of Political Economy - 75 - August - pp.352-65. Booth, A.L., Snower, D.J. (1996), Acquiring Skills, Cambridge University Press, Cambridge Burtless, G., Orr, L. (1986), "Are classical experiments needed for manpower policy?", Journal of Human Resources, Vol. 21 No.4, pp.606-39. Chapman, P. (1993), The Economics of Training, Harvester, Manchester. Chinapah, V., & Miron, G. (1990). Evaluating educational programmes and projects: Holistic and practical considerations. Socio-economic studies, No. 15. Paris: UNESCO. Deacon, B (2000) ‘Eastern European Welfare States: the impact of the politics of globalization.’, Journal of Social European Policy, Vol 10 (2) pp. 146-161 Doeringer, P.B., Piore, M.J. (1985), Internal Labour Markets and Manpower Analysis, Sharp Armonk, New York, NY, . European Union (1999), Socio-economic Programs Evaluation, Collection MEANS, Luxembourg, European Union (2002) Impact Evaluation of the European Employment Strategy: Technical Analysis. http://ec.europa.eu/employment_social/employment_strategy/eval/papers/technical_anal ysis_complete.pdf Ferrera, M,; Hemerijck A.; Rhodes M. (2000) The Future of Social Europe. Recasting Work and Welfare in the New Economy Oeiras (Portugal) Celta Editora. Fits-zen, J (2000) The ROI of Human Capital: Measuring the Economic Value of Employee performance. AMACON. Gramlich E. (1990) A Guide to Cost Benefit Analysis. Second Edition, Prentice Hall. New Jersey Heckman J.; Hotz J. (1989) Choosing Among Alternative Nonexperimental Methods for Estimating the Impact of Social Programs: The Case of Manpower Training Journal of the American Statistical Association, Vol. 84, No. 408 (Dec., 1989), pp. 862-874 Ignjatovic, M, Svetlik, I. (2003) European HRM clusters. EKK Toim., Autumn 2003, no. 17, str. 25-39. Kaplan R.; Norton D (1996) The Balance Scorecard. Harvard. Katz, E, Ziderman, A (1990), "Investment in general training: the role of information and labour mobility", Economic Journal, Vol. 100 No.403, pp.1147-58. 16 Keynes, J.(1990) Teoria Geral do Emprego, do Juro e da Moeda - São Paulo - Editora Atlas - Translated from the 1973 edition of the McMillan Press. Kirkpatrick D. (1998) Evaluating Training Programs: The Four Levels. (2nd edition) San Francisco. Berrett-Kohler. LaLonde, R. (1995), "The promise of public sponsored training programs", Journal of Economic Perspectives, Vol. 9 No.2, pp.149-68. Mc Lean G. (2005) Examining Approaches to HR evaluation; Strategic Human Resources Volume 4, Issue 2, January – February. Mincer, J.(1958) Investment in On Job Training and Personal Incomes Distribution Journal of Political Economy - 66 - 4 - pp.281-392. Mincer, J. (1962) On-the-job training: costs, returns, and implications. - The Journal of Political Economy - volume LXX - 5 - part 2 - Supplement - October - pp.50-80. Mincer, J. (1974) Schooling, Experience and Earnings – NBER – Columbia University Press – New York. Nonaka I.; Takeuchi H. (1995) The Knowledge Creating Company – Oxford University Press. OECD (2006) Education at glance. Paris. Oi, W. (1962) Labour as a quasi-fixed factor – The Journal of Political Economy - 70 – pp.538-55. Orr, L. (Eds) (1996) Does Training for the Disadvantaged Work? Evidence From the National JTPA Study, Urban Institute Press, Washington, DC, . Paul J.(1989) La relation formation - emploi: un défi pour l' économie, Economica Paris. Shackleton, J (1995) Training and Unemployment in Western Europe and the United States -Edward Elgar , Aldershot, England. Schultz T. (1971) Investment in Human Capital: The Role of Education and of Research, New York: Free Press. Spence, A M. 1973. Job Market Signaling. Quarterly Journal of Economics, 87(3), pp. 355-374. Stevens, M. (1996) Transferable training and poaching externalities - in Alison Booth and Denis Snower editors - Acquiring skills: market failures, their symptoms, and policies responses - CEPR - Cambridge University Press - pp.19-40. Stiglitz, J (1975). The Theory of "Screening," Education, and the Distribution of Income. American Economic Review, 65(3), pp. 283-300. Tomé E. (2001) The Evaluation of Vocational Training: A Comparative Analysis Journal of European Industrial Training, volume 25 number 7 pags. 380-88 2001 Tomé E. (2005) Evaluation Methods in Social Policy – Bits – Boletin informativo de trabajo social – nº 8. UNESCO (2006) Participation in formal technical and vocational education training programmes worldwide. An initial statistical study. UNEVOC International Centre for Technical and Vocational Education and Training. Bonn, Germany. United Nations (2008) Human Development Report 2007/8- Fighting climate change: Human solidarity in a divided world. New York. Internet websites: Eurostat (2007) Labour market policy expenditure in active measures by type. 17