Table of contents - EGAD Project

advertisement
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Questionnaire for Evaluation
of an Engineering Program
Submitted by:
Queen’s University at Kingston, Ontario
Mechanical Engineering
August 31, 2011
Canadian Engineering Accreditation Board
1100 – 180 Elgin Street, Ottawa, ON K2P 2K3
Tel.: (613) 232-2474 / Fax: (613) 230-5759
ceab@engineerscanada.ca
1
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Table of contents
Table of contents .......................................................................................................... 2
Glossary of terms .......................................................................................................... 2
3. Accreditation criteria ................................................................................................. 3
3.0 Continual improvement .......................................................................................... 3
Summarize the continual improvement process: .................................................................... 3
3.1
Graduate attributes ........................................................................................... 4
Appendix 3.1 Graduate Attributes.................................................................................... 20
Glossary of terms
Accreditation Units (AU) are defined on an hourly basis for an activity which is granted academic
credit and for which the associated number of hours corresponds to the actual contact time of that
activity between the student and the faculty members, or designated alternates, responsible for
delivering the program:

one hour of lecture (corresponding to 50 minutes of activity) = 1 AU

one hour of laboratory or scheduled tutorial = 0.5 AU
This definition is applicable to most lectures and periods of laboratory or tutorial work. Classes of
other than the nominal 50-minute duration are treated proportionally. In assessing the time assigned to
determine the AU of various components of the curriculum, the actual instruction time exclusive of
final examinations should be used.
Assessment tool: sources of data on student learning.
Illustrative examples: design report, presentation, essay, examination, standardized exam,
oral examination, observed behaviour, focus group, survey, etc.
Curriculum content: This need not mean an entire course dedicated to specific material; for example,
it may include separate units within an array of courses which address the material.
Curriculum map: a plotted representation (often in the form of a table) that shows the relationship
between learning experiences (e.g. courses, co-ops, co-curricular activities), instructional and
assessment methods, and intended learning for each aspect of a given program so that the
relationships and connections among all the elements are easily seen.
Illustrative Example: If a program identifies three indicators to demonstrate the graduate
attribute Lifelong Learning, as described above, a table could be used to show which learning
experiences (e.g. courses) are used to develop abilities and assess indicators
Graduate Attribute: generic characteristics, specified by the CEAB, expected to be exhibited by
graduates of Canadian engineering schools
Indicator: descriptors of what students must do to be considered competent in the attribute; the
measurable and pre-determined standards used to evaluate learning
2
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Illustrative example: Criterion 3.1.12 requires that students possess the attribute Lifelong Learning. A
program might consider that the indicators required to demonstrate that students possess this attribute
are:



Critically evaluates procured information for authority, currency, and objectivity.
Describes professional and academic societies in the discipline and how new knowledge
enters discipline.
Identifies resources and professional associations that address student’s own ongoing
professional development.
Minimum path: The set of courses, including common-core, program compulsory, option compulsory
and elective courses which provide the least number of AU within each Accreditation Board curriculum
content category.
Modern engineering tools: This refers to tools such as equipment, processes, codes of practice,
software, simulation packages, etc. that are considered essential for the given discipline.
“Qualified” Accreditation Units: Curriculum content delivered by faculty members that meet the
Accreditation Board accreditation licensure requirements. Engineering licensure is examined only for
courses that include engineering science and/or engineering design curriculum content. Please see the
Statement of interpretation on licensure expectations and requirements for further information.
Weakest-link principle: All options in the program are examined. Following the principle that a
program is only as strong as its “weakest link”, a program is accredited only if all such variations meet
the criteria.
3. Accreditation criteria
The following sections describe the measures used by the Accreditation Board to evaluate Canadian
engineering programs for the purpose of accreditation. The number indicated beside each criterion
corresponds with the Accreditation Board’s Accreditation Criteria and Procedures for 2010.
3.0 Continual improvement
Summarize the continual improvement process:
In 2009 the Faculty of Engineering and Applied Science began developing a process for graduate
attribute assessment by forming working groups to establish measurable and meaningful
indicators of graduate attributes appropriate to first year, middle years, and at graduation. In
the 2009-2010 academic year the Faculty piloted a process for attribute assessment within
some faculty-wide first year courses. In the summer of 2010 the processes were reviewed and
expanded, and in the 2010-2011 academic year the process of attribute assessment was
broadened to involve include faculty wide first and second year courses, and the fourth year
capstone design, communications and project management courses MECH 460/464.
3
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
For this cycle attribute assessment was a separate process from the remainder of accreditation
preparation. As a result the detailed response to Criterion 3.0 and 3.1 are contained in a
separate document in Appendix 3.1, which details the processes followed faculty wide and
common to all programs for the 2010-2011 academic year. It lists the indicators measured,
where they were measured, a description of how the curriculum contributes to the
development of the attributes, a description of the student activities used to measure the
attributes, data collected from student work and surveys, an evaluation of the data, and
proposals for program improvement. Based on evaluation of the data from student work and
surveys, revisions are planned as follows:





The existence and importance of attributes for engineering practice will be
communicated and used more extensively, and linked to learning objectives in courses.
Some of the attributes appear to be poorly understood by students, and this will be
addressed
At the first year level, the program is being revised in the areas including making
effective arguments, evaluating complex problem solutions against objectives, written
communications, and evaluating information. Grader calibration is being enhanced to
reduce variation between graders, and some indicators will be revised to reflect
student ability.
At the second year level, changes are being made to the faculty-wide communications
and economics courses. More emphasis will be placed on summarizing important
information clearly and concisely, effectively participating in informal small group
discussions, and on risk assessment and project planning.
A 4-year sequence of courses in engineering design and practice is partway through
implementation, which will support development of professional skill attributes, and
provide a venue for developing and assessing all attributes in broad integrative
experiences, like team projects, that emulate professional practice.
The outcomes assessed in the Mechanical and Materials Engineering capstones MECH
460/464 (See Below) will be harmonized with the new outcomes developed in the
faculty wide courses. A project this summer has collected a list of over 1000 outcomes
being assessed, but not yet recorded, in our current offerings. Analysis of this data will
allow a more complete picture of our program in combination with the faculty wide
curriculum rolling out this fall.
Because the process of attribute assessment generally started in faculty wide courses, the
faculty-wide data and analysis is more detailed and should be viewed as representing an
example of process that will augment our graduating year assessment in the coming years. The
faculty plans to continue assessing student attributes throughout the four-year program for the
purpose of program improvement. As curriculum changes occur over the next two years, the
program evaluation and improvement process will be expanded to include more comprehensive
coverage of all attributes at the first and graduating years of the program.
3.1
Graduate attributes
The higher education institution must demonstrate that the graduates of a program possess the
attributes under the following headings. The attributes will be interpreted in the context of candidates
4
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
at the time of graduation. It is recognized that graduates will continue to build on the foundations that
their engineering education has provided.
Engineering programs are expected to continually improve. There must be processes in place that
demonstrate that program outcomes are being assessed in the context of these attributes, and that the
results are applied to the further development of the program.
Response to criterion 3.1:
This section presents only data specific to the Mechanical and Materials Engineering (MME)
program and is additional to the information on process and faculty wide curriculum presented
in Appendix 3.1. The details presented in sections 3.1.1 through 3.1.12 cover only the
assessments made in the capstone courses MECH 460/464 and do not include the faculty wide
information found in the appendix.
MME Program Curriculum Map
In the summer of 2011 a project in MME collected a list of over 1000 outcomes based on a
retrospective survey of assessments actually applied in MME program courses during the
2010/2011 year. For each assessment (report, exam question, problem set) in each course, the
marks assigned were associated with an outcome or outcomes such as “Applies Bernoulli’s
equation to a pipe flow” or “Composes clear and logical arguments”. Those outcomes were
categorized by CEAB Attribute, CDIO Syllabus Topic, and Bloom Verb for knowledge
characterization, among other details. This analysis is not yet complete, however the
curriculum map table below lists courses where teaching and assessment has been carried out
in each of the Attribute areas. The collected outcomes will be used as explicit indicators to
gather data on student performance starting in 2011/2012, with initial concentration on
courses with strong relationships to attributes 2 through 12 to capture performance in areas
other than the Engineering Knowledge Base.
Attribute
Developed
Assessed
Knowledge base
Extensively
Extensively
5
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
problem analysis
Extensively
APSC-100, MECH-396,397,398,399
Investigation
APSC-100, MECH
215/396/397/398/399
(not this year)
Design
APSC-100, MECH 212/323
APSC-100,MECH-323,212,460, Math272
Engineering
tools
APSC -100, 142, 161, MECH
323
APSC 161, Math 272, MECH 213, MECH 323
Communication
APSC-100, 291, 292, MECH
464
APSC-100, 291,292, MECH213,396,397,398,399,464
Individual and
team work
APSC-100, MECH 212
APSC-100, MECH 212
Impact of
engineering on
society
APSC-100, 151
APSC-100, 151
Ethics and
equity
APSC-100
APSC-100
Professionalism
APSC-100, MECH 212
APSC-100, MECH 212, 460,464
Economics and
project
management
APSC-100, 221
APSC-100, 221
Lifelong learning
APSC-100, 291, 292
APSC-100, 291, 292, MECH 464
Graduating year assessment
In 2010-2011, the final year class was assessed on the basis of performance in the jointly
offered MECH 460/464 capstone design, communications and project management courses.
Students were assessed both on a team basis (Group Assignment in the legend) and as
individuals (Assignment) for outcomes taken from the faculty wide list in development and
from the CDIO Syllabus where additional coverage was required. All outcomes were assessed
directly by regular or adjunct faculty members who are Professional Engineers.
Data were accumulated for each outcome during the marking process and stored in the Moodle
course management system. The charts that follow were completed based on Moodle records
at the end of the academic year. The data for all students in both courses was downloaded,
aligned to provide a complete record for each student, and then anonymized for privacy. The
resulting frequencies were accumulated over all students and plotted for each outcome in each
assignment. The overall data set includes 96 students, the majority of whom participated in
both courses. The data does not capture the performance of students taking the APSC 480
Multidisciplinary Design course.
Each of the FEAS outcomes were ranked on a scale of:




Not Demonstrated, which included instances where the outcome was not addressed
Marginal
Meets Expectations
Exceeds Expectations
6
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
CDIO Syllabus outcomes were ranked on an Engineering Practice scale of:





Resubmit: The work is unsatisfactory and must be revised.
File Only: The work is sufficient to provide a record for the file.
Internal Circulation: I would show this to my boss.
Almost Client Ready: Only a few minor tweaks are required.
Client Ready: The work is ready for submission to an outside client with no revisions.
All of the outcomes are satisfactory except Resubmit and Not Demonstrated. File Only or
Marginal is below expectations, but sufficient to pass the course.
The full numerical data is available in the file MECH460—F-464—FW Combined Outcomes
Anonymized.xlsx available as part of the on-site exhibits.
The following outcomes were used as indicators in the graduating year for MME students. The
faculty wide indicators have been supplemented by several CDIO indicators as needed. Each
indicator (or outcome) was explicitly assessed as part of the MECH 460/464 capstone project
course evaluation by project faculty supervisors.
7
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Sec.
3.02
3.02
Attribute
Code
Indicator
Course Deliverable
Problem analysis
3.02-GY1
Identifies problem, known and unknown
information, uncertainties, and biases
Final Report
3.02-GY2
Creates process for solving problem
including justified approximations and
assumptions
Final Report
Final Report
Problem analysis
3.02
Problem analysis
3.02-GY3
Selects and applies appropriate model and
analysis to solve problems
3.02
Problem analysis
3.02-GY4
Evaluates validity of results and model for
error, uncertainty
Final Report
3.04-GY1
Identify problem and constraints including
health and safety risks, applicable
standards, economic, environmental,
cultural and societal considerations
Proposal Report
3.04-GY2
Applies appropriate knowledge,
judgement, and design tools, in creating
and analyzing conceptual design solutions
to select best concept
Proposal Report
3.04-GY3
Creates and tests simulations, models,
and/or prototypes at various points in
design with complexity appropriate to
design stage
Final Report
3.04-GY4
Assesses design performance based on
requirements, yield, reliability, and/or
safety as appropriate
Final Report
3.04-GY5
Identifies possibilities for further
improvement and conducts design review
to evaluate performance of the overall
process.
Final Report
Final Report
3.04
3.04
3.04
3.04
3.04
Design
Design
Design
Design
Design
3.05
Engineering tools
3.05-GY1
Evaluates techniques, resources, and tools
to identify their limitations with respect
to needs
3.05
Engineering tools
3.05-GY2
Applies appropriate techniques, tools, and
processes to accomplish a task
Final Report
3.05
Engineering tools
3.05-GY3
Evaluates appropriateness of results from
several engineering techniques and tools
Final Report
Demonstrates capacity for initiative and
technical or team leadership while
respecting others' roles
Meeting Leader,
Memo and Minutes /
Supervisor Assessment
of Individual
Contribution
3.06
Teamwork
3.06-GY1
3.06
Teamwork
3.06-GY2
Evaluates team effectiveness and plans for
improvements
Not Explicitly
Assessed, although
students completed
peer evaluations at
mid term and the end
of term.
3.07
Communications
3.07-GY1
Writes and revises documents using
appropriate discipline-specific conventions
Proposal Report
3.07
Communications
3.07-GY2
Demonstrates accurate use of technical
vocabulary
Proposal Report
3.07
Communications
3.07-GY3
Demonstrates confidence in formal and
Meeting Leader,
8
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
informal oral communications
Memo and Minutes/
Supervisor Assessment
of Individual
Contribution
3.07
Communications
3.07-GY4
Uses appropriate referencing to cite
previous work
Proposal Report
3.07
Communications
3.07-GY5
Uses graphics to explain, interpret, and
assess information
Proposal Report
CDIO 3.2.3
Compose meeting minutes and progress
memo
Meeting Leader,
Memo and Minutes
CDIO 3.2.6
Demonstrate answering questions
effectively
Meeting Leader,
Memo and Minutes /
Proposal Presentation
or Final Poster
Presentation
Final Report
3.07
3.07
Communications
Communications
3.08
Professionalism
3.08-GY1
Integrates standards, codes of practice,
and legal and regulatory factors into
decision-making processes (as
appropriate)
3.08
Professionalism
3.08-GY2
Demonstrates professional bearing
Meeting Leader,
Memo and Minutes
3.09
Impact of
engineering
3.09-GY1
Considers economic, social, and
environmental factors and/or impacts in
decisions
Final Report
3.09
Impact of
engineering
3.09-GY2
Evaluates trade-offs among goals and
concepts
Proposal Report
3.09
Impact of
engineering
3.09-GY3
Explains the societal, enterprise, and/or
technical context of the system
Proposal Report
3.11
Economics and
project
management
CDIO 4.3.4
Apply project Control for cost,
performance, and schedule in the
estimation and allocation of resources
considering risks and alternatives.
(typically includes a Gantt Chart and Work
plan)
Proposal Report
3.12
Lifelong learning
3.12-GY1
Critically evaluates procured information
for authority, currency, and objectivity.
Proposal Report
Demonstrate skills of self-education
Supervisor Assessment
of Individual
Contribution
3.12
Lifelong learning
3.12-GY2
Overall Student performance on outcome measures is strong, with 90% meeting or exceeding
expectations on individual outcomes. This performance is consistent across the attributes
assessed. The remaining students showed marginal outcomes in these areas, sufficient to meet
our threshold requirements to pass the course. These marginal outcomes were not strongly
clustered in individual students or project groups so that in aggregate, all but eight students
met our expectations and those eight met expectations in half or more of the outcomes.
Outcomes of Not Demonstrated appear under several attributes. These instances were
investigated with faculty supervisors. They are generally a result of perceptions by the students
and faculty supervisor that addressing that outcome was not essential to that particular
project. For example, a team designing an on-board braking system for an SAE Mini-Baja
9
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
vehicle would have had to make a big reach to include substantial elements of impact on
society and the environment. This is the first application of these outcomes in our final year
projects and many project scopes were defined before the assessment scheme was finalized.
We will address this in the future by ensuring that the scope of each project includes elements
representative of the full outcomes list, in combination with assessing some outcomes, e.g. 10.
Ethics and Equity in other parts of the curriculum or on deliverables not directly tied to
capstone project reports.
Students show considerable progress in communications abilities over time in comparison to the
similar cohort assessed this year in APSC 291/292. Anecdotal observation also indicates
substantial progress in design, team, and project skills from their performance in the MECH 212
design course. Repeated experience over multiple years in contexts based on engineering
practice and design in teams develops the skills necessary for engineering practice. This is not
surprising.
Our students are strongest in the outcomes based on presentations in various forms ranging
from informal meetings to poster sessions and formal slide presentations. Marginal outcomes
appear in more substantial numbers in both the technical and communications elements of
their proposals and final reports, although the overall quality is still very high.
We are in the process of expanding the outcomes assessment to other courses to allow us to
collect more comprehensive data on student performance with respect to the graduate
attributes. This will be especially useful where all students are working on activities with the
same objectives and scope, all marked by the same faculty member. Individual faculty
variation adds considerable noise to the assessments on final year projects.
10
This chart shows the outcomes from the combined MECH 460/464 results grouped by assignments.
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
11
This chart shows the outcomes from the combined MECH 460/464 results grouped by CEAB Attribute.
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
12
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Mechanical Engineering Program specific development
Mechanical and Materials Engineering has been an active partner in the development of the
faculty wide initiatives summarized above and detailed in Appendix 3.1, and has been actively
revising and improving curricula towards design and engineering practice based goals since at
least the early nineties. We pioneered development of curriculum and prototype facilities
eventually incorporated in the Integrated Learning Initiative and the Integrated Learning
Centre, including:



MECH 212 design course in second year that incorporates two project activities and
design build within a context of engineering practice. The new APSC 200 design course
is largely modelled on MECH 212’s success.
The ILC Active Learning Centre was prototyped with found materials on the third floor
of Jackson Hall in support of MECH 212.
The ILC Plaza Labs were prototyped on the first floor of Jackson Hall to support more
open-ended lab activities in MECH 215.
Mechanical and Materials Engineering at Queen’s was one of the founding members of the CDIO
Initiative in Engineering Education (cdio.org) which now includes more than 70 member
institutions around the world. We hosted the First International CDIO Conference in 2005 and
we continue active participation and one of our faculty, Rick Sellens, holds a seat on the CDIO
Council.
CDIO Standards are closely aligned with the Washington Accord and with national and
international accreditation requirements based on them, such as ABET (a-k) and the CEAB
Graduate Attributes. Mapping by Cloutier, Hugo and Sellens1 shows that the highly detailed
CDIO Syllabus is a superset of the topics required to meet the CEAB Graduate Attributes and
shows how programs may select from the Syllabus to meet the Attributes. Many of the Queen’s
faculty wide outcomes listed above are adapted from CDIO and additional outcomes based on
CDIO have been used in evaluating some aspects of our program not covered by the faculty
wide outcomes.
In 2005 we completed an initial assessment of our curriculum against CDIO Standards and
Syllabus, including a mapping of our courses to the relevant areas of the Syllabus on an
Introduce, Teach, Utilize basis, and completion of stakeholder surveys. The stakeholder surveys
showed consistency with others internationally and strong agreement among faculty, students,
alumni and industry partners. A major result of that assessment was a refocusing of design
activities in all years, especially capstone (see section 3.3.4.4 of the MME Questionnaire), to
focus on industrially relevant design activities with practical consequences in the real world. In
addition, many faculty members reworked elements of their engineering science courses to
deliver materials in the context of real systems and engineering practice following CDIO
Standard 1.
1
Guy Cloutier, Ron Hugo and Rick Sellens, Mapping the relationship between the CDIO Syllabus and the CEAB
Graduate Attributes: An Update, paper 132, 7th International CDIO Conference 2011, Copenhagen, Denmark, June
20-23, 2011.
13
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
In the summer of 2011, we are gathering a detailed list of what outcomes were actually
assessed across our core curriculum in 2010/2011 for comparison against the earlier mapping
and to determine which changes have actually stuck and been carried through to student
assessment. These outcomes from individual courses are being categorized in terms of both the
CDIO Syllabus and the CEAB Attributes to provide further evidence that they are parallel
systems with very similar objectives. Once the mapping is complete we will have data on how
many of the AU within our core program relate to each of the different topics in the Syllabus
and to each of the different CEAB Graduate Attributes. Where possible, the outcomes will be
monitored in Moodle to allow us to collect comprehensive data on our student performance,
and support a further process of continuous improvement.
3.1.1 A knowledge base for engineering
Response to criterion 3.1.1:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program and indicators have been identified as part of our summer 2011
curriculum project. These indicators will be evaluated in 2011/2012.
The data available from student performance in technical courses indicate that students are
meeting or exceeding our expectations for this attribute as a requirement of graduation.
The results from the 2011/2012 assessments will be combined with the faculty wide data from
Appendix 3.1 and follow the process defined there to improve coverage and student
performance as part of our ongoing curriculum monitoring process under the CDIO Standards.
3.1.2 Problem analysis
Response to criterion 3.1.2:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program and indicators were collected as part of the assessment process in
MECH 460/464 as detailed above. Indicators evaluated for this attribute were:
“Identifies problem, known and unknown information, uncertainties, and biases” was evaluated
in Final Report by the Faculty Supervisor.
“Creates process for solving problem including justified approximations and assumptions” was
evaluated in Final Report by the Faculty Supervisor.
“Selects and applies appropriate model and analysis to solve problems” was evaluated in Final
Report by the Faculty Supervisor.
“Evaluates validity of results and model for error, uncertainty” was evaluated in Final Report
by the Faculty Supervisor.
14
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
The data indicate that more than 90% of students are meeting or exceeding our expectations
for this attribute at the fourth year level. However, about 25% of students did not adequately
evaluate the validity and uncertainty of their solutions.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards. Emphasis will be added to their
reporting requirements to increase their focus on validity of results.
3.1.3 Investigation
Response to criterion 3.1.3:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, particularly in MECH 215/396/397/398 lab courses. Indicators
have been identified as part of our summer 2011 curriculum project and will be applied to the
lab course in 2011/2012.
The data collected by direct performance assessment will indicate if students are meeting or
exceeding our expectations for this attribute.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards.
3.1.4 Design
Response to criterion 3.1.4:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, including APSC 100, MECH 212/323/460, and indicators were
collected as part of the assessment process in MECH 460/464 as detailed above. Indicators
evaluated for this attribute were:
“Identify problem and constraints including health and safety risks, applicable standards,
economic, environmental, cultural and societal considerations” was evaluated in Proposal
Report by the Faculty Supervisor.
“Applies appropriate knowledge, judgement, and design tools, in creating and analyzing
conceptual design solutions to select best concept” was evaluated in Proposal Report by the
Faculty Supervisor.
“Creates and tests simulations, models, and/or prototypes at various points in design with
complexity appropriate to design stage” was evaluated in Final Report by the Faculty
Supervisor.
“Assesses design performance based on requirements, yield, reliability, and/or safety as
appropriate” was evaluated in Final Report by the Faculty Supervisor.
“Identifies possibilities for further improvement and conducts design review to evaluate
performance of the overall process” was evaluated in Final Report by the Faculty Supervisor.
15
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
The data indicate that over 90% of students are meeting or exceeding our expectations for this
attribute at the fourth year level. About 12% of students do not adequately transfer knowledge
from the technical courses to their design project.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards. Knowledge application will be
improved through our increasing integration of design activities into engineering science
courses under CDIO.
3.1.5 Use of engineering tools
Response to criterion 3.1.5:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, including APSC 100/142/161, MATH 272, and MECH 212/323 and
indicators were collected as part of the assessment process in MECH 460/464 as detailed
above. Indicators evaluated for this attribute were:
“Evaluates techniques, resources, and tools to identify their limitations with respect to needs”
was evaluated in Final Report by the Faculty Supervisor.
“Applies appropriate techniques, tools, and processes to accomplish a task” was evaluated in
Final Report by the Faculty Supervisor.
“Evaluates appropriateness of results from several engineering techniques and tools” was
evaluated in Final Report by the Faculty Supervisor.
The data indicate that more than 90% of students are meeting or exceeding our expectations
for this attribute at the fourth year level. Many students did not explicitly evaluate the
appropriateness of tools for the application, typically in instances where the appropriateness
was obvious, as in applying a finite element package to stress analysis in a component.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards.
3.1.6 Individual and team work
Response to criterion 3.1.6:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, including APSC 100 and MECH 212/460/464, and indicators were
collected as part of the assessment process in MECH 460/464 as detailed above. Indicators
evaluated for this attribute were:
“Demonstrates capacity for initiative and technical or team leadership while respecting others'
roles” was evaluated in Meeting Leader, Memo and Minutes / Supervisor Assessment of
Individual Contribution by the Faculty Supervisor.
16
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
“Evaluates team effectiveness and plans for improvements” was not explicitly assessed,
although students completed peer evaluations at mid term and the end of term.
The data indicate that students are meeting or exceeding our expectations for this attribute at
the fourth year level. This is unsurprising given their teaming experiences in APSC 100 followed
by the MECH 212 design course which provides the model for the new faculty wide APSC 200
course.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards.
3.1.7 Communication skills
Response to criterion 3.1.7:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, including APSC 100/291/292 and MECH 464, and indicators were
collected as part of the assessment process in MECH 460/464 as detailed above. Indicators
evaluated for this attribute were:
“Writes and revises documents using appropriate discipline-specific conventions” was
evaluated in Proposal Report by the Faculty Supervisor.
“Demonstrates accurate use of technical vocabulary” was evaluated in Proposal Report by the
Faculty Supervisor.
“Demonstrates confidence in formal and informal oral communications” was evaluated in
Meeting Leader, Memo and Minutes / Supervisor Assessment of Individual Contribution by the
Faculty Supervisor.
“Uses appropriate referencing to cite previous work” was evaluated in Proposal Report by the
Faculty Supervisor.
“Uses graphics to explain, interpret, and assess information” was evaluated in Proposal Report
by the Faculty Supervisor.
“Compose meeting minutes and progress memo” was evaluated in Meeting Leader, Memo and
Minutes by the Faculty Supervisor.
“Demonstrate answering questions effectively” was evaluated in Meeting Leader, Memo and
Minutes / Proposal Presentation or Final Poster Presentation by the Faculty Supervisor.
The data indicate that all students are meeting or exceeding our expectations for oral
communications at the fourth year level, with the exception of a few students who are
marginal in responding to questions during oral presentations. Less than 10% of our students
still have some difficulty meeting expectations in the details of referencing, vocabulary, and
discipline specific conventions in their writing, but are otherwise strong in their written
communications.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
17
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
curriculum monitoring process under the CDIO Standards. More emphasis will be place on the
details in MECH 464 lectures and additional detailed material will be incorporated into the new
APSC 293 communications course paired with the faculty wide APSC 200 design project course,
which should produce improvements in future years.
3.1.8 Professionalism
Response to criterion 3.1.8:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, including APSC 100/151 and MECH 212, and indicators were
collected as part of the assessment process in MECH 460/464 as detailed above. Indicators
evaluated for this attribute were:
“Integrates standards, codes of practice, and legal and regulatory factors into decision-making
processes (as appropriate)” was evaluated in Final Report by the Faculty Supervisor.
“Demonstrates professional bearing was evaluated” in Meeting Leader, Memo and Minutes by
the Faculty Supervisor.
The data indicate that more than 90% of students are meeting or exceeding our expectations
for this attribute at the fourth year level. Some design projects fall in areas outside the realms
of established standards and codes of practice.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards. Greater attention will be given to
including elements of standards within all design projects.
3.1.9 Impact of engineering on society and the environment
Response to criterion 3.1.9:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, including APSC 100/151, and indicators were collected as part of
the assessment process in MECH 460/464 as detailed above. Indicators evaluated for this
attribute were:
“Considers economic, social, and environmental factors and/or impacts in decisions” was
evaluated in Final Report by the Faculty Supervisor.
“Evaluates trade-offs among goals and concepts” was evaluated in Proposal Report by the
Faculty Supervisor.
“Explains the societal, enterprise, and/or technical context of the system” was evaluated in
Proposal Report by the Faculty Supervisor.
The data indicate that over 95% of students are meeting or exceeding our expectations for this
attribute at the fourth year level. In some design projects their assessment of context
indicated minimal social or environmental impacts and thus no opportunity to consider
mitigation of those impacts.
18
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards. Greater attention will be given to
including elements of social and environmental impact within the scope of all design projects.
3.1.10 Ethics and equity
Response to criterion 3.1.10:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, primarily in APSC 100, but not explicitly assessed in the MME
program outside the faculty wide courses.
Anecdotal sources suggest that most students are meeting or exceeding our expectations for
this attribute, however empirical data is required.
We will review the coverage in the faculty wide data from Appendix 3.1 and the new second
year curriculum coming on line this fall and follow the process defined there to improve
coverage and student performance as part of our ongoing curriculum monitoring process under
the CDIO Standards.
3.1.11 Economics and project management
Response to criterion 3.1.11:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, particularly in APSC 100/221 and MECH 464, and indicators were
collected as part of the assessment process in MECH 460/464 as detailed above. Indicators
evaluated for this attribute were:
“Apply project Control for cost, performance, and schedule in the estimation and allocation of
resources considering risks and alternatives. (Typically includes a Gantt Chart and Work plan)”
was evaluated in Proposal Report by the Faculty Supervisor.
The data indicate that over 90% of students are meeting or exceeding our expectations for this
attribute at the fourth year level.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards.
3.1.12 Life-long learning
Response to criterion 3.1.12:
Details for faculty wide courses are presented in Appendix 3.1. This attribute is developed in
various parts of the program, particularly in APSC 100/291/292 and MECH 464, and indicators
were collected as part of the assessment process in MECH 460/464 as detailed above.
Indicators evaluated for this attribute were:
19
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
“Critically evaluates procured information for authority, currency, and objectivity” was
evaluated in Proposal Report by the Faculty Supervisor.
“Demonstrate skills of self-education” was evaluated in Supervisor Assessment of Individual
Contribution by the Faculty Supervisor.
The data indicate that over 90% of students are meeting or exceeding our expectations for this
attribute at the fourth year level.
These results will be combined with the faculty wide data from Appendix 3.1 and follow the
process defined there to improve coverage and student performance as part of our ongoing
curriculum monitoring process under the CDIO Standards.
Appendix 3.1 Graduate Attributes
Appendix 3.1 Contents
Appendix 3.1 Contents................................................................................................................................ 20
Section 3.0 – Continual improvement process ........................................................................................... 22
Section 3.1 Part A: Demonstration of student attributes ........................................................................... 24
Purpose and outcomes ........................................................................................................................... 25
Curriculum mapping ............................................................................................................................... 29
Assessment and evaluation .................................................................................................................... 33
Faculty-wide First year ........................................................................................................................ 33
Faculty-wide middle year assessment: Communications ................................................................... 69
Faculty-wide Middle year assessment: Economics ............................................................................. 75
Student input on Graduate attributes ................................................................................................ 77
Program development informed by data ............................................................................................... 82
Faculty wide development .................................................................................................................. 82
Section 3.1 Part B: Future plans for graduate attribute assessment .......................................................... 83
Appendix 3.1A: Leveled indicators.............................................................................................................. 85
Appendix 3.1B: Student surveys ................................................................................................................. 91
Appendix 3.1C: Definitions and Guidelines................................................................................................. 96
Criterion 3.1 ............................................................................................................................................ 96
Supplementary guidelines for Criterion 3.1 ............................................... Error! Bookmark not defined.
Definitions .............................................................................................. Error! Bookmark not defined.
20
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
21
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Section 3.0 – Continual improvement process
Engineering programs are expected to continually improve. There must be processes in place that
demonstrate that program outcomes are being assessed in the context of the graduate attributes, and
that the results are applied to the further development of the program. Summarize the continual
improvement process.
The Faculty of Engineering and Applied Science is committed to continual program improvement that
has historically focused on many of the attributes in Criterion 3.1. In the mid 90’s the faculty began an
initiative, labelled Integrated Learning, to promote integration of theory and practice, and concepts
from academia and industry. The faculty saw a need to enhance development of professional skills and
an understanding of the engineering profession in the first year program. Informed by data from pilot
projects and an awareness of outcomes-based assessment being introduced in the USA via EC2000, a
committee of faculty, administration, and industry advisors designed new facilities and curriculum to
developing competence in areas now covered by the CEAB graduate attributes. This led to the
introduction of two new team and project-based first year courses focusing on problem analysis, design,
investigation, engineering tools, teamwork, communications, professionalism, ethics and equity, and
lifelong learning.2
Over the past 10 years these courses have seen regular re-development based on student feedback and
assessment3. Previously existing courses have also been going through a development process; a course
on earth systems has been developing a stronger emphasis on the impact of engineering on society and
the environment, and courses on engineering graphical communications and computer programming
include open-ended design projects.
With the introduction of the graduate attribute assessment requirements the faculty has begun
formalizing an approach to program development informed by measurable and meaningful data.
In the 2009-2010 academic year the Faculty piloted a process for attribute assessment within some first
year courses. At the end of the year the processes were reviewed and expanded to cover graduating
year courses for the 2010-2011 academic year. The process being developed is shown in Figure 1 and
follows other accepted processes for outcomes assessment.4
In 2009 the Faculty created seven working groups to establish measurable and meaningful indicators
appropriate to first year, middle years, and at graduation. Although current accreditation regulations
2
See J. McCowan, C. Knapper, An Integrated and Comprehensive Approach to Engineering Curricula, (Three parts),
Int. J. Eng. Ed, Vol. 18, No. 6, 2002.
3
See A. Topper, L. Clapham, From Experiments to Experimentation; A New Philosophy for First Year Laboratories,
C2E2 2001, Victoria; B. Frank, J. Mason, Impact of Peer-Managed Project-Based Learning in First Year Engineering,
ASEE 2008 General Conference, Pittsburgh, PA
4
P. Wolf, New Directions for Teaching and Learning, Volume 2007, Issue 112, pp. 15-20
22
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
require assessment only at graduation, the faculty is assessing attribute development in earlier years for
several reasons:
1. The ability to apply the results of graduate attribute assessment to the further development of
the program is improved when data is available for earlier years.
2. Creating expectations by years greatly improves the efficiency and quality of curriculum design
3. The first year of the engineering program is common to all engineering students and is
administered by the faculty office. It is efficient to benchmark all incoming students.
4. Most of the professional skill attributes are specifically targeted in the first two years of the
current program and not in upper years. The curriculum is being developed to ensure these
attributes continue to be developed in upper years.
Figure 1 - Current graduate attribute assessment process
Working groups were composed of representatives from departments across the faculty, subject experts
in areas required by the graduate attributes (library science, economics, communications, etc.) and
students. An educational developer was consulted throughout the process. The groups used several
resources including the CDIO syllabus and resources from other engineering accreditation bodies. The
23
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
requirements from the Ontario Council of Academic Vice Presidents Guidelines for University
Undergraduate Degree Learning Expectations (UDLEs) were also used.5
1.
Depth and Breadth of Knowledge
2.
Knowledge of Methodologies
3.
Application of Knowledge
4.
Communication Skills
5.
Awareness of Limits of Knowledge
6.
Autonomy and Professional Capacity
Appendix 3.1A shows the indicators created by these groups in a single table.
At the beginning of the 2010-2011 academic year, a set of indicators from Appendix 3.1A was selected
for assessment in the 2010-2011 academic year based on the current state of the curriculum and ability
to assess in a timely manner. The indicators were mapped to some faculty-wide courses, including the
common first year, and faculty-wide middle year courses in economics and communication, and to
program specific courses. Appropriate assessment tools were determined by individual course
instructors. Finally, grading resources like rubrics were created for each of the course assignments that
were used to assess the indicators.
In a parallel process, a survey of instructors in the faculty was conducted to look at the alignment
between the curriculum and the attributes.
After the data was gathered and collated, statistical analysis was performed and provided to the faculty
and programs to interpret and create an improvement plan.
Section 3.1 Part A: Demonstration of student attributes
Instructions for criterion 3.1
A. Describe the processes that are being or are planned to be used to demonstrate that
students possess the attributes required by criteria 3.1.1 to 3.1.12, and that the results are
being applied to the further development of the program. The description should include:
a) a set of indicators that describe specific abilities expected of students to demonstrate
each attribute
b) where attributes are developed and assessed within the program, such as via courses,
co-curricular activities, co-op positions (e.g., a curriculum map). Ensure that courses
described are easily identified in the course descriptions presented on the Course
Information Sheets.
5
See, e.g. http://www.cou.on.ca/related-sites/the-ontario-universities-council-on-qualityassura/policies/pdfs/quality-assurance-framework---guide-may-2011.aspx
24
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
c) how the indicators were or will be assessed. This could be based on assessment tools
that include, but are not limited to, reports, oral presentations, observed laboratory
performance, assignments, exams, standardized tests, student/alumni surveys, peer
reviews, self-assessments, and work term experience/reports which may be used in
whole or in part. For example, evidence may be demonstrated via a single question
embedded on an exam or in the entirety of one or more of the assessment tools. Provide
samples of work or other documents deemed to exceed, meet, and fall below the
expected level of performance. How the samples of work or other documents are used to
assess indicators should be clearly explained.
d) evaluation of the data collected including analysis of student performance relative to
program expectations
e) discussion of how the results will be used to further develop the program
f)
a description of the ongoing process used by the program to assess and develop the
program as described in (a)-(e) above.
The response to part A above must specifically address each graduate attribute in criteria
3.1.1 to 3.1.12.
Purpose and outcomes
The Mission statement for the Faculty of Engineering and Applied Science is:
We educate engineering students for leadership and citizenship in a global society through high quality,
technically rigorous engineering programs.
The Faculty’s perspective was used to develop indicators faculty-wide for first year, middle years, and
graduating year; these are shown in Appendix A to give a perspective on the general approach being
taken. The tables below show the indicators that were assessed in the 2010-2011 academic year, and
the courses in which they were assessed.
25
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Faculty-wide first year indicators
Sec.
Attribute
Code
3.01
Knowledge base
3.01-FY1
3.01
Knowledge base
3.01-FY2
3.01
Knowledge base
3.01-FY3
3.02
Problem analysis
3.02-FY1
3.02
Problem analysis
3.02-FY2
3.02
Problem analysis
3.02-FY3
3.02
Problem analysis
3.02-FY4
3.04
Design
3.04-FY1
3.04
3.04
Design
Design
3.04-FY2
3.04-FY3
3.04
Design
3.04-FY4
3.04
Design
3.04-FY5
3.04
Design
3.04-FY6
3.04
Design
3.04-FY7
3.06
Teamwork
3.06-FY1
3.06
Teamwork
3.06-FY2
3.06
Teamwork
3.06-FY3
3.06
Teamwork
3.06-FY7
3.07
Communications
3.07-FY3
3.08
3.08-FY2
3.09
Professionalism
Impact of
engineering
3.10
Ethics and equity
3.10-FY2
3.10
Ethics and equity
3.10-FY3
3.09-FY4
Indicator
Create mathematical descriptions or
expressions to model a real-world problem
Select and describe appropriate tools to solve
mathematical problems that arise from
modeling a real-world problem
Use solution to mathematical problems to
inform the real-world problem that gave rise to
it.
Identifies known and unknown information,
uncertainties, and biases when presented a
complex ill-structured problem
Creates process for solving problem including
justified approximations and assumptions
Selects and applies appropriate quantitative
model and analysis to solve problems
Evaluates validity of results and model for error,
uncertainty
Adapts general design process to design
system, component, or process to solve openended complex problem.
Accurately identifies significance and nature of
a complex, open-ended problem
Identifies customer and user needs
Gathers and uses information from appropriate
sources, including applicable standards,
patents, regulations as appropriate.
Produces a variety of potential design solutions
suited to meet functional specifications
Performs systematic evaluations of the degree
to which several design concept options meet
project criteria
Compares the design solution against the
problem objective
Recognizes a variety of working and learning
preferences
Applies principles of conflict management to
resolve team issues
Assumes responsibility for own work and
participates equitably
Exercises initiative and contributes to team
goal-setting
Summarizes and paraphrases written work
accurately with appropriate citations
Demonstrates punctuality, responsibility and
appropriate communication etiquette
Incorporates sustainability considerations in
decision making (societal and ecological)
Identifies and iterates items from the
professional codes of conduct
Describes ethical issues and how they affect
the individual, the company and the public
Course
APSC-171
APSC-171
APSC-171
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
26
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
3.11
Economics
3.11-FY1
3.11
Economics
3.11-FY2
3.11
Economics
3.11-FY3
3.12
Lifelong learning
3.12-FY1
3.12
3.12
Lifelong learning
Lifelong learning
3.12-FY2
3.12-FY4
3.12
Lifelong learning
3.12-FY5
3.12
Lifelong learning
3.12-FY6
Plans and efficiently manages time and money
Establishes appropriate project scope, after
consultation with client, based on available
resources.
Plan maps out project with clear milestones
and delegation.
Uses information effectively, ethically, and
legally to accomplish a specific purpose,
including clear attribution of Information
sources.
Identifies a specific learning need or knowledge
gap.
Describes own learning style
Identifies appropriate technical literature and
other information sources to meet a need
Critically evaluates the procured information for
authority, currency, and objectivity.
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100
Faculty-wide middle year indicators
Two attributes, communications and engineering economics, were assessed within the middle years of
the program as there are faculty-wide courses targeting these attributes within the current curriculum.
APSC-291 and -292 are second year courses in communications, each with 12 AU. APSC-221 is a course
in engineering economics and business taken by MME students, usually in their fourth year.
Indicators for these courses are shown below. Note that the code for middle year communications
follows the CDIO Syllabus convention, in contrast to other indictor codes. These codes are coursespecific.
Sec.
Attribute
3.07
Communications
3.07
Communications
Code
2.2.2
2.3.3a
3.07
Indicator
Course
Apply information search and
identification, with proper citations
Identify all relevant factors and the
dominant factors in the system
APSC291/292
Communications
APSC291/292
2.3.3b
3.07
Communications
3.07
Communications
APSC291/292
2.4.4
2.4.5a
Identify and Prioritize relevant factors
Compose logical arguments based on
supporting evidence
Assess the extent of one's presentation
abilities, and one's responsibility for selfimprovement to overcome important
weaknesses
APSC291/292
APSC291/292
27
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
3.07
Communications
2.4.5b
3.07
Communications
3.07
Communications
3.07
Communications
3.07
Communications
3.1.2
3.2.1a
3.2.2a
Communications
3.2.2b
3.07
Communications
3.07
Communications
3.2.2c
3.2.2d
3.07
Communications
3.07
Communications
3.2.3a
3.2.3b
3.07
Communications
3.07
Communications
3.2.3c
3.2.3d
3.07
Communications
3.2.3e
3.07
Communications
3.07
Communications
3.2.3f
3.2.3g
3.07
Communications
3.07
Communications
3.2.5
3.2.6a
3.07
3.07
Communications
Communications
APSC291/292
APSC291/292
APSC291/292
APSC291/292
3.2.1b
3.07
Assess the extent of one's team
communications skills, and one's
responsibility for self-improvement to
overcome important weaknesses
Demonstrate effective communication in a
small team (active listening, collaboration,
providing and obtaining information)
Demonstrate a communications strategy
that considers the needs and character of
the audience
3.2.6b
3.2.6c
Identify and Summarize key content
Compose with the appropriate structure
and relationship amongst ideas
Create logical, persuasive arguments in
your cover letter
Demonstrate conciseness, crispness,
precision and clarity of language
Show relevant, credible, accurate
supporting evidence in your resume
Compose clear, precise, sequential
description of events or instructions
Demonstrate writing with coherence and
flow
Demonstrate writing with correct spelling,
punctuation and grammar
Use appropriate formatting for a memo,
including to: and cc:
Use appropriate formatting for a short
report, including headings and figure
captions
Use appropriate formatting for an
executive summary
Use prescribed formatting for a letter and
resume
Demonstrate construction of block
diagrams
Prepare a presentation with appropriate
language, style, timing and flow
Prepare and deliver a presentation with
appropriate language, style, timing and
flow
Prepare presentation materials with
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
APSC291/292
28
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
3.07
Communications
3.2.6d
3.07
Communications
4.3.1
3.11MY1
3.11
Economics
3.11
Economics
3.11
Economics
3.11
Economics
3.11MY2
3.11MY3
3.11MY4
3.11
Economics
3.11MY5
appropriate language, style, timing and
flow
Use appropriate nonverbal
communications (gestures, eye contact,
poise)
Apply the language/format of goals and
requirements
Gather appropriate information, categorize it,
and determine the economic attractiveness of
an engineering project
Measure the financial impact of risks that are
associated with the engineering project and
consider the risk and return relationship as a
component of determining economic
attractiveness
Plans a project from scoping to completion,
including scheduling
Plans for the management of risks and changes
required for project
Describes new enterprise formation, business
planning and the economic attractiveness of the
enterprise.
APSC291/292
APSC291/292
APSC-221
APSC-221
APSC-221
APSC-221
APSC-221
Graduating year
MME specific information for the graduating year is detailed in section 3.1 of the main body of the
questionnaire.
Curriculum mapping
In addition to mapping indicators to courses as shown above, a survey of instructors in the faculty was
conducted to look at the alignment between the curriculum and the graduate attributes.6 Each course
instructor was asked to respond to a set of questions about each course they teach. The questions were
as follows:
a. What methods of instruction do you use in your course? (respondents select from list
provided)
b. What methods of assessment are used in your course? (respondents select from list
provided)
6
Queen`s was generously allowed to pilot the CurricKit curriculum mapping tool created at the University of
Guelph for this survey. Information is at: http://currickit.wikispaces.com/
29
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
c. Which of the Engineering Graduate Attributes are developed in your course?
(respondents select from list provided)
d. What level of complexity/depth is expected for each of the identified Engineering
Graduate Attribute? (respondents select one of three levels)
e. Please specify the ways that each grad attribute is Taught in your course. (respondents
select from the list in part (a))
f. Please specify the ways that each grad attribute is Assessed in your course.
(respondents select from list in part (b))
g. What is the mark and workload distribution over the duration of the semester?
h. Are optional assignments or optional weighting of assignments made available to your
students? If so, please specify.
i. Do you have any general comments?
The plot below shows the results across all of engineering. Knowledge base and problem analysis are the
most commonly developed and assessed attributes, and economics, ethics and equity, lifelong learning,
and professionalism are the least. This distribution agrees with student perceptions as discussed later in
this report.
30
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
The plot below shows the expected complexity for the attributes as reported by course instructors.
Again, knowledge base and problem analysis attributes have the greatest number of courses reporting
an expectation for work-ready performance.
Within only faculty-wide courses (APSC), the distribution is shown below. The courses include all of the
core first year courses, second and third year communications and economics courses. The prevalence
of professional skill development is under-represented in the plot below, as many of the skills are
developed in a course that is equivalent to three typical term-length courses (APSC-100, which has 132
AU).
31
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
A rough mapping of the attributes to courses in the existing faculty wide curriculum is shown below.
Details of the program specific course mapping for Mechanical and Materials Engineering is provided in
Section 3.1 of the main body of the questionnaire. Section 3.1B discusses proposals for curriculum
development that will expand the development and assessment of graduate attributes.
Attribute
Knowledge base
Problem analysis
Investigation
Design
Engineering tools
Communication
Individual and team work
Impact of engineering on society
Ethics and equity
Professionalism
Economics and project
management
Lifelong learning
Developed
Extensively
Extensively
APSC-100,
APSC-100,
APSC -100, 142, 161
APSC-100, 291, 292
APSC-100
APSC-100, 151
APSC-100
APSC-100
APSC-100, 221
Assessed
(not this year)
APSC-100
(not this year)
APSC-100,
(not this year)
APSC-100,291
APSC-100
APSC-100
APSC-100
APSC-100
APSC-100,221
APSC-100
APSC-100
32
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Assessment and evaluation
Faculty-wide First year
Historically many of the knowledge and skills now targeted by the graduate attributes have been
developed in the common first year of the engineering program. As a result, all 11 of the 12 attributes
were assessed in two first year course courses, in addition to the assessment process in the graduating
year. Mathematical knowledge base for engineering was assessed in APSC-171, Calculus I, and ten of the
professional skill attributes were assessed in APSC-100, Engineering Practice. Details and sample work
may be found in the example material provided for each course.
The sections below show the indicators, assessment tools, and data gathered from these courses in the
2010-2011 academic year.
Knowledge base for Engineering
Three specific indicators were assessed in the context of introductory mathematical knowledge. These
reflect the three key course objectives of the introductory calculus course APSC-171, Calculus I.
Code
3.01-FY1
3.01-FY2
3.01-FY3
Indicator
Create and/or select mathematical descriptions or expressions for simple
real-world problems involving rates of change and processes of
accumulation
Course
APSC-171
Select and describe appropriate tools to solve mathematical problems
that arise from modeling a real-world problem
Use solution to mathematical problems to inform the real-world problem
that gave rise to it.
APSC-171
APSC-171
These indicators are also appropriate for the problem analysis graduate attribute, but since they are
specific to mathematics they are listed under the knowledge attribute. The final exam in the course was
designed to allow these three indicators to be assessed individually by specific questions or subquestions on two tests and the final exam. A copy of these tests is shown in the course material for
APSC-171. The material below shows results from specific questions on tests and the exam that were
developed to evaluate student performance on these indicators.7
7
The material in this section was adapted from a report prepared by the course coordinator of APSC-171, Prof. L.
Jonker.
33
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Indicator 1: Create and/or select mathematical descriptions or expressions for simple real-world
problems involving rates of change and processes of accumulation
The context for this question involved calculating intersection of two trajectories. One of the important
ways in which mathematics connects to the "real" world is through its applications to geometric space
and dynamics. This connection was explored especially during the first few weeks of the course and was
tested especially on the first test. Thus, the third and fourth pages of the first test consist of a question
that gives students the equations of the trajectories of two particles and then asks them to determine
whether the particles will collide, and (separately) whether their trajectories cross, and at what point of
time the velocity of the second particle is in a horizontal direction. This material is difficult for the
students, especially for those who do not come from Ontario high schools, where the curriculum
includes a unit on linear algebra and vectors. The marks for these pages show this, for they were quite
low, averaging 3.8 out of 6 (63%).
Figure 2
Indicator 2: Select and describe appropriate tools to solve mathematical problems that arise from
modeling a real-world problem
The context for this was differentiation similar to that required in the high school curriculum. Generally,
we find that students are well prepared to negotiate the second indictor, and the test and exam
questions that test this are done very well by most. For example, in Test 1, Question 1, students were
asked to calculate the derivatives of mathematical functions involving logarithms, exponents, powers.
These problems test students' knowledge of the various rules of differentiation, as well as their
knowledge of logarithm and exponential laws, and of the derivatives of trigonometric functions. Some of
that (particularly the differentiation rules and the laws of exponents and logarithms) was on their high
school curriculum. It shows, as the average mark (out of 11) was 8.9 (81%).
34
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Figure 3
Another question on the first test that was purely focused on mathematical knowledge and problem
analysis was Question 7. The first part of this question tested students' understanding of inverse
trigonometric functions, while the second part tested their understanding of implicit differential
equations. This question was done less well because both were entirely new topics to the bulk of the
students, and the first one, particularly, involved some subtlety that students tend to miss. The average
mark on this question was 3.3 out of 6 (55%).
35
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Figure 4
Indicator 3: Use solution to mathematical problems to inform the real-world problem that gave rise to
it.
One of the challenges in teaching calculus is that, whereas students know how to differentiate when
they come out of high school, their understanding of what a derivative is in practice is not as secure as it
might be. We consider one of the main goals of the course to be that they should gain a better, more
practical, sense of what a derivative represents. One of the ways in which we do this is by spending
quite a bit of time asking students to interpret differential equations (mostly without teaching them to
solve a differential equation!) and to translate statements of physical laws into mathematics (inevitably
resulting in a differential equation). One of the questions on Test 1 (Question 6) probes this
understanding, and asks students to do this sort of translation, together with a linear approximation
(the essence of the derivative):
Question 6 on Test 1: This problem concerns atmospheric pressure near the earth. A rough model is
that:
“The rate of change in pressure with height is proportional to the pressure at that height."
(a) Write down the mathematical equation that expresses this relationship. Be sure and label all your
variables and constants as clearly as possible, and do not use any variables that you do not need. Finally,
state whether the constants in your equation are positive or negative.
(b) The constant of proportionality in part (a) has an absolute value of approximately 0.124 /km. At a
height of 5:6 km above sea level, the pressure is approximately 0.5 atm (atmospheres). What is the rate
of change of pressure at this height?
36
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
(c) Use your answer from (b) and linear approximation to estimate the pressure at a height of 6 km
above sea level.
The average on this question was 4.9 out of 7 (70%). The histogram for the marks indicates that some
students continue to find this exercise difficult.
Problem analysis, Design, Communications, etc.
Ten of the professional skill attributes were assessed in APSC-100, Engineering Practice (the
Investigation attribute was not assessed this year, but will be in future years). These were assessed in
integrated assignments which covered multiple attributes. The tables below show the deliverables by
which indicators were assessed. Details about each deliverable expectation are described in the APSC100 course material.
APSC-100 is divided into three modules: Module 1 focuses on complex problem solving and numeric
computation (MATLAB is used as the tool); Module 2 focuses on investigation and design of
experiments; Module 3 centres around a semester long client-based team design project. Modules 1
and 3 were used to assess attributes this year.
Indicators used in APSC100 Module 1 (Fall and winter)
Course Deliverable
Assignment: Learning styles and professional
communication
Handed in: Sept 2010
Assignment: Problem solving and modeling activity 1
Indicators
FEAS - 3.12-FY4: Describes learning style
FEAS - 3.02-FY1: Identify problem
FEAS - 3.02-FY2: Create process
37
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
(PSM1: Mersey Venture Elevator Failure)
Handed in: Oct 2010
Assignment: Problem solving and modeling activity 2
(PSM2: Water filter)
Handed in: Nov 2010
Assignment: Problem solving and modeling activity 4
(PSM4: nicotine diffusion)
Handed in: Feb 2010
FEAS - 3.02-FY3: Select model
FEAS - 3.02-FY4: Evaluate solution
FEAS - 3.10-FY3: Describes ethical issues
FEAS - 3.02-FY1: Identify problem
FEAS - 3.02-FY2: Create process
FEAS - 3.02-FY3: Select model
FEAS - 3.02-FY4: Evaluate solution
FEAS - 3.10-FY3: Describes ethical issues
FEAS - 3.02-FY2: Create process
FEAS - 3.02-FY3: Select model
FEAS - 3.02-FY4: Evalute solution
Nicotine Diffusion - Ethical dilemma
Indicators used in APSC100 Module 3 (Winter)
In Module 3 student deliverables were assessed by three groups of people: each project had a faculty
sponsor (FS), an upper year student serving as a project manager (PM), and independent academic
assistants who evaluated written and oral communications. In the data below, the evaluators are
indicated by FS for faculty sponsors, PM for project managers, and AA for academic assistants.
Course Deliverables
Midterm Individual Assessment
Description
Assessment on Students’ individual performance to the middle
of semester
- Conflict management 3.06-FY2
- Responsibility 3.06-FY3
- Initiative 3.06-FY7
- Professional behavior 3.08-FY2
Final Individual Assessment
Assessment on Students’ individual performance to the end of
semester
- Conflict management 3.06-FY2
- Responsibility 3.06-FY3
- Initiative 3.06-FY7
- Professional behavior 3.08-FY2
Team Assignment 1: Project
definition (TA1)
The first report done by groups. Contents focus on the project
definition and team organization.
- 3.04-FY2: Identify problem
- 3.04-FY3: Identify needs
- 3.12-FY2: Identifies learning need
- 3.12-FY5: Identifies appropriate literature
- 3.06-FY1: Learning preferences
Team Assignment 2: Project scope
and information (TA2)
A brief report outlines team project plan and summarizes the
information research.
- 3.11-FY3: Project plan
- 3.11-FY2: Project scope
- 3.07-FY3: Summarizes accurately
- 3.12-FY6: Evaluates information
- 3.12-FY1: Ethical information use
- APSC-100 Comm: Conciseness
38
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Proposal Report Draft
In order to achieve a formal and professional proposal report, a
draft is required. Rubrics here focus on the grammar,
formatting and reference.
- 3.07-FY3: Summarizes accurately
- APSC100 Comm: Purpose and style
- APSC100 Comm: Coherence and Format
- APSC100 Comm: Graphical communications
- APSC100 Comm: Correctness
Proposal Report
The formal/final version of proposal report. Rubrics here are
different than the one above.
- 3.04-FY2: Identify problem
- 3.11-FY3: Project plan
- 3.04-FY5: Potential solutions
- 3.04-FY6: Decision making
- 3.07-FY3: Summarizes accurately
- APSC100 Econ: Proposal economic analysis
- APSC100M3 Proposal: Overall proposal
Proposal Presentation
A presentation on the proposal of the project.
- APSC-100 Oral Comm: Organization
- APSC-100 Oral Comm: Content
- APSC-100 Oral Comm: Slides
- APSC-100 Oral Comm: Speakers
- APSC-100 Oral Comm: Purpose and style
Final Report
The Formal final report which summarizes students’ term
efforts
- 3.04-FY4: Gathers info
- 3.11-FY1: Manage time and money
- 3.04-FY1: Uses process
- 3.09-FY4: Sustainability in decisions
- 3.04-FY7: Compares solution
- APSC 100 Final Report – Implementation
- APSC 100 Final Report - Overall Written Communication
Final Presentation
A presentation on the final solution/implementation of the
project
- APSC-100 Oral Comm: Organization
- APSC-100 Oral Comm: Content
- APSC-100 Oral Comm: Slides
- APSC-100 Oral Comm: Speakers
- APSC-100 Oral Comm: Purpose and style
The Table below summarizes which assignments were used to assess each indicators across the entire
course.
39
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
X
X
X
X
X
X
X
X
Final Report
Proposal Report
Proposal Report Draft
Team Assignment 1: Project
Definition (TA1)
Team Assignment 2: Project Scope
and Information (TA2)
Final Individual Assessment
Midterm Individual Assessment
Module 3
Problem Solving and Modeling
Activity 4 (PSM4)
Problem Solving and Modeling
Activity 2 (PSM2)
Indicators
FEAS – 3.02-FY1
FEAS – 3.02-FY2
FEAS – 3.02-FY3
FEAS – 3.02-FY4
FEAS – 3.04-FY1
FEAS – 3.04-FY2
FEAS – 3.04-FY3
FEAS – 3.04-FY4
FEAS – 3.04-FY5
FEAS – 3.04-FY6
FEAS – 3.04-FY7
FEAS – 3.06-FY1
FEAS – 3.06-FY2
FEAS – 3.06-FY3
FEAS – 3.06-FY7
FEAS – 3.07-FY3
FEAS – 3.08-FY2
FEAS – 3.09-FY4
FEAS – 3.10-FY3
FEAS – 3.11-FY1
FEAS – 3.11-FY2
FEAS – 3.11-FY3
FEAS – 3.12-FY1
FEAS – 3.12-FY2
FEAS – 3.12-FY4
FEAS – 3.12-FY5
FEAS – 3.12-FY6
Assignment: Learning Styles and
Professional Communication
Problem Solving and Modeling
Activity 1 (PSM1)
Module 1
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
40
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Some indicators were added specifically to APSC-100, beyond the faculty-wide indicators created and
shown in Appendix 3.1A. Most indicators were scored on a four-level rubric, with a “1” corresponding to
expectation not being demonstrated, a “2” meaning marginal (threshold) performance, “3” for meeting
expectations (target performance), and 4 for exceeding expectations. The rubrics were given to
students when the assignments were created.
Each deliverable, by students or team, was graded using the rubric. The total grade for the deliverable
was used to compute the mark on the assignment, and the score for each indicator was stored for the
purpose of analyzing students’ ability to meet expectations. In the next section student performance is
described.
Sample rubrics for some of the deliverables are shown below. More detail is shown in the set of APSC100 course materials. Rubric descriptors for each indicator on each deliverable are shown in the section
with student scores.
41
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
APSC-100 (2010 -2011) MEA1: Mersey Venture Elevator Failure
Information summary
3.02-FY1: Identify problem
1
(not demonstrated)
Information not identified
properly, no information,
or information copied
from assignment.
2
(marginal)
Some important
information or biases not
identified, or
trivial/incorrect
information included.
3
(meets expectations)
Identifies known and
unknown information,
uncertainties, and biases.
4
(outstanding)
Meets expectations plus:
No or inadequate
process
Process identified
misses some important
factors; some
assumptions left
unidentified or
unjustified.
Creates justified process
for solving problem,
suppored by information.
Meets expectations plus:
No analysis, or
model/analysis selected
is inappropriate
Model selected, some
errors in analysis or
inappropriate
assumptions.
Selects and applies
appropriate quantitative
model and MATLAB
analysis to solve problems,
using reasonable
approximations and
assumptions.
Meets expectations plus:
No evaluation of
solution.
Superficial evaluation of
solution
Evaluates validity of results
and model for error,
uncertainty
Meets expectations plus:
Makes supported
recommendations to
prevent future failures;
describes implications to
stakeholders.
Meets expectations plus:
Mark
(/4)
Includes information from
authoritative sources to
inform process, model,
and conclusions.
/4
Proposed process
3.02-FY2: Create process
Comprehensive process
model; comparison with
other possible approaches.
/4
Model
3.02-FY3: Select model
Model results and
conclusions
3.02-FY4: Evalute solution
Recommendations
3.10-FY3: Describes
ethical issues
No or trivial
recommendations
Some important
recommendations and
implications missed.
Authoritative research
used to defend
assumptions and
approximations made
/4
Evaluates conclusions and
presents potential
improvements.
Information, model, and
conclusions used to make
recommendations
supported by authoritative
information.
/4
/4
42
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
APSC-100 (2010 -2011) MEA2: Clay water filter
Information summary
3.02-FY1: Identify problem
Proposed process
3.02-FY2: Create process
Analysis
3.02-FY3: Select model
Model results and
conclusions
1
(not demonstrated)
Information not
identified properly, no
information, or
information copied
from assignment.
2
(marginal)
Some important
information or biases
not identified, or
trivial/incorrect
information included.
3
(meets expectations)
Identifies known and
unknown information,
uncertainties, and biases.
No or inadequate
process
Process identified
misses some important
factors; some
assumptions left
unidentified or
unjustified.
Creates justified process
for solving problem,
including concept map,
that includes T,S,E,F
factors.
Model selected, some
errors in analysis or
inappropriate
assumptions.
Uses analysis to make
decision on multiple
factors using reasonable
approximations and
assumptions.
Meets expectations plus:
Authoritative research used
to defend assumptions and
approximations made
Evaluates validity of
results and model for
error, impact of
uncertainty.
Meets expectations plus:
Evaluates conclusions and
presents potential
improvements.
/4
Describes potential side
effects of design.
Meets expectations plus:
Potential side effects
informed by model, analysis,
and conclusions, supported
by authoritative information.
/4
No analysis, or
model/analysis
selected is
inappropriate
No evaluation of
solution.
Superficial evaluation of
solution
3.02-FY4: Evalute solution
Side effects
3.10-FY3: Describes ethical
issues
No or trivial
recommendations
Some important
recommendations and
implications missed.
4
(outstanding)
Meets expectations plus:
Includes information from
authoritative sources to
inform process, model, and
conclusions.
Mark
(/4)
/4
Meets expectations plus:
Comprehensive process
model; comparison with
other possible approaches.
/4
/4
43
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
APSC-100 (2010-2011)
Purpose and style
Coherence and
Format
Sequence,
transitions,
formatting
1
(not demonstrated)
Unclear purpose. Tone and
style inappropriate.
Audience needs not met.
Poorly organized; rambling,
lacks unity; Inconsistent
writing/formatting; lacks
transitions; many gaps or
redundancies. Handwritten
material.
Graphical
communications
Figures and tables not
related to text. Improper
referencing of figures. List of
tables/figures missing.
Correctness
Numerous grammar errors;
disorganized sentence
structure; inappropriate
units or significant digits;
acronyms not defined
Erroneous or incomplete
conclusions drawn from
sources, or inappropriate
quality or quantity of
background information.
Information usage
3.07-FY3: Summarizes
accurately
Overall:
Written Communications Rubric
2
(marginal)
Some aspects of presentation
don’t achieve purpose.
Organization sometimes
unclear; some unsuitable
sections, minor gaps or
redundancies, minor formatting
problems; some wordy
expressions . "This is/There is/It
is" …
Some figures and tables not
discussed in text. Figure/table
captions missing. Incomplete/
inconsistent list of tables/
figures.
Some grammar/ punctuation
errors; spelling; incomplete
sentences.
Sources summarized with minor
misconceptions.
3
(meets expectations)
Clear purpose is met.
Formal tone and style
appropriate to audience
Organized, appropriate
sections, uniformly and
correctly formatted; little
irrelevant information.
4
(outstanding)
Professional tone and style.
Authoritative and
convincing
Focused, logically organized;
skillful and varied
transitions. Professionally
formatted. No irrelevant
information
Figures and tables referred
to in text, captioned.
Appropriate lists of
figures/tables.
Figures and tables
professionally formatted,
integrated into text,
complementing text
Relatively free of grammar/
punctuation errors; units
specified; acronyms and
symbols defined
No typos or grammatical
errors; attractively and
professionally formatted.
Summarizes and
paraphrases appropriate
sources accurately with
appropriate citations
Multiple authoritative,
objective, reliable sources
used; cited and formatted
properly
Mark
(/4)
/4
/4
/4
/4
/4
/
44
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
APSC-100 (2010-2011)
Purpose & Style
Organization
Sequence;
transitions. Methods
to emphasize content
(illustrations,
repetition)
Content
Oral Communications Rubric
1
(not demonstrated)
Unclear purpose. Tone
and style inappropriate.
Audience needs not
met.
Insufficient content;
poorly organized;
rambling; not
understandable
Minimal content that
does not demonstrate
appropriate progress on
project.
Slides
Mechanics,
acronyms, units,
graphics
Numerous typos and
grammatical errors;
unprofessional and
inconsistent formatting;
figures and tables
unclear or unrelated
Speakers
Volume, speed,
cadence, English
usage; questions.
Use of aids
Hard to understand/hear
the speaker; speaking
far too fast/slow;
questions not
satisfactorily answered;
reading from notes
Team Number:
2
(marginal)
Some aspects of
presentation don’t
achieve purpose.
3
(meets expectations)
Clear purpose is met
Formal tone and style
appropriate to audience
Organization sometimes
unclear; some
unsuitable sections;
disconnected. Some
useful information
missing.
Generally appropriate
content but some
information missing.
Organized; appropriate
information; generally
smooth transitions
between topics
Some typos; some
inconsistency in
formatting; some
acronyms and symbols
not defined;
inappropriate use of
figures/tables
Generally understood
with some volume, pitch
and speed issues; some
questions answered
incorrectly; reading off of
slides
Date:
Evaluator:
4
(outstanding)
Professional tone and
style
Authoritative and
convincing
Slides logical, easy to
follow; detailed,
impressive depth of
thought; excellent
transitions between topics
Appropriate content
covering project
background, scope,
relevant information,
design options, and work
completed.
Relatively free of typos and
grammatical errors
Slides uniformly formatted
Acronyms and symbols
defined; appropriate
graphics
Excellent content and
evidence of detailed
thought and analysis.
Understandable,
appropriate volume, pitch
and speed; questions
generally answered
correctly; aids used
properly
Professional variation in
volume, pitch, speed; very
knowledgeable and clear
when handling questions;
aids used very well
Mark
(/4)
Slides free of typos and
grammatical errors; slides
uniformly formatted and
professional; material and
graphics support main
ideas
Overall:
45
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
APSC-100 Module 3: Final team evaluation
1
(not demonstrated)
2
(marginal)
3
(meets expectations)
4
(outstanding)
Mark
Information Management No significant
information used, not
3.04-FY4: Gathers info
cited; blatant
plagiarism.
Insufficient usage; improper
citations.
Gathers and uses information
from appropriate sources,
including applicable standards,
patents, regulations as
appropriate, with proper citations
Uses information from multiple
authoritative, objective, reliable
sources; cited and formatted
properly
Project Management
No useful timeline or
budget described;
poorly managed
project; safety issues
Poor timeline or budget;
infrequent meetings; minor
safety problems
Plans and efficiently manages
time and money; team effectively
used meetings; safety
considerations are clear
Efficient, excellent project plan
presented; detailed budget;
potential risks foreseen and
mitigated
No discussion of
design process.
Generic design process
described.
Describes design process used to
design system, component, or
process to solve open-ended
complex problem.
Comprehensive design process
described, with appropriate
iterations and revisions based on
project progress
/4
No consideration of
these factors.
Factors mentioned but no
clear evidence of impact on
decision making.
Incorporated appropriate social,
environmental, and financial
factors in decision making
Well-reasoned analysis of these
factors, with risks mitigated where
possible
/4
Implementation
Insufficient
implementation
Sufficient implementation but
some opportunities not
taken, or feedback at
proposal not incorporated in
implementation
Appropriate effort, analysis,
Outstanding implementation
and/or construction demonstrated
to implement product, process, or
system
/4
Evaluation
No evaluation of
design solution
Some factors missed in
evaluating design solution
Compares the design solution
against the project objectives and
functional specifications, providing
qualitative evaluation where
appropriate
/4
Poorly constructed
report
Some organization
problems, minor formatting
problems, redundancy,
spelling grammar/errors
Report achieves goal using formal Professional tone, convincing
tone, properly formatted,
argument, authoritative, skillful
concisely written, appropriate use transitions
of figures, few spelling/grammar
errors
3.11-FY1: Manage time
and money
Design process
3.04-FY1: Uses process
Social, environmental,
and financial factors
/4
/4
3.09-FY4: Sustainability in
decisions
3.04-FY7: Compares
solution
Written communication
Comprehensive evaluation of
design solution, with welldefended recommendations for
future work or implementation
/4
46
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Attribute: Problem analysis
The plots below show student scores on indicators of problem analysis. In these plots, and plots below,
the text in parentheses shows the abbreviation of the deliverable by which the indicator was assessed.
Abbreviations are shown in the list of indicators above. The table below shows the levelled descriptors
used by graders to score student work on these indicators.
The problem analysis attribute was identified as one where students fail to meet expectations more
than others at the first year level. The indicators on which more than 20% of the students fell below the
target expectation were:


3.02-FY1: Identify problem (twice)
3.02-FY4: Evaluate solution (three times)
Figure 5
1 - Not Demonstrated
3.02 - FY1: Identifies known and Information not
unknown information,
identified properly,
uncertainties, and biases when no information, or
presented a complex illinformation copied
structured problem
from assignment
Threshold
Target
2 - Marginal
3 - Meets Expectations
Some important
information or biases not
Identifies known and unknown
identified, or
information, uncertainties,
trivial/incorrect information and biases
included
3.02 - FY2: Creates process for
solving problem including
justified approximations and
assumptions
Process identified misses
Creates justified process for
some important factors;
solving problem, suppored by
some assumptions left
information.
unidentified or unjustified.
Meets expectations PLUS:
Comprehensive process model;
comparison with other possible
approaches
No analysis, or
3.02 - FY3: Selects and applies
model/analysis
appropriate quantitative model
selected is
and analysis to solve problems
inappropriate
Selects and applies approriate
quantitative model and
Model selected; some
MATLAB analysis to solve
errors in analysis or
problems, using reasonable
inappropriate assumptions
approximations and
assumptions
Meets expectations PLUS:
Authoritative research used to
defend assumptions and
approximations made
3.02 - FY4: Evaluates validity of
results and model for error,
uncertainty
Superficial evaluation of
solution
No or inadequate
process
No evaluation of
solution
Evaluates validity of results
and model for error,
uncertainty
4 - Outstanding
Meets expectations PLUS:
Includes information from
authoritative sources to inform
process, model, and conclusions
Meets expectations PLUS:
Evaluates conclusions and
presents potential improvements
47
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Design
Plots of performance scores on design indicators are shown below. Generally performance on these was
acceptable. However, given the large percentage of students scored as “Exceeding expectations”, the
wording of the indicators needs to be reviewed to ensure that the expectations are appropriate.
Figure 6
1 - Not Demonstrated
3.04 - FY1: Adapts general
design process to design
No discussion of
system, component, or process design process.
to solve open-ended complex
problem.
Project definition is
3.04 - FY2: Accurately identifies
loosely defined.
significance and nature of a
Difficulty narrowing
complex, open-ended problem
goal of project
3.04 - FY3: Identifies customer
and user needs
3.04 - FY4: Gathers and uses
information from appropriate
sources, including applicable
standards, patents, regulations
as appropriate.
Threshold
2 - Marginal
Generic design process
described.
Target
3 - Meets Expectations
4 - Outstanding
Describes design process used
to design system, component,
or process to solve openended complex problem.
Comprehensive design process
described, with appropriate
iterations and revisions based on
project progress
Thoroughly identifies and
Accurately identifies and
Repeats project description
describes the presented problem
describes the presented
as presented by client or
while being aware of human
problem building on the client
course
factors, resources constraints and
requirements
client needs
Statement is multidimensional
No description of
Identifies customer and user
Identifies some customer
showing constraints and potential
customer or user
needs; elaborates and explain
and user needs.
strengths. Is aware of potential
needs
how it will impact the project
bias from client
Gathers and uses information
No significant
from appropriate sources,
Uses information from multiple
information used, not Insufficient usage;
including applicable standards, authoritative, objective, reliable
cited; blatant
improper citations.
patents, regulations as
sources; cited and formatted
plagiarism.
appropriate, with proper
properly
citations
48
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Figure 7
1 - Not Demonstrated
3.04 - FY5: Produces a variety of
No description of
potential design solutions
potential design
suited to meet functional
solutions
specifications
3.04 - FY6: Performs systematic
evaluations of the degree to
which several design concept
options meet project criteria
No
comparison/analysis
of potential design
solutions
3.04 - FY7: Compares the design
No evaluation of
solution against the problem
design solution
objective
Threshold
2 - Marginal
Target
3 - Meets Expectations
4 - Outstanding
Some critical considerations Produces a variety of potential Applies creative approaches to
missed in generating
design solutions suited to
identify and develop alternative
potential design solutions meet functional specifications concepts and procedures
Contrasts potential design
solutions without analysis
Performs systematic
evaluations of the degree to
which several design concept
options meet project criteria
Uses technical knowledge,
mathematical models,
appropriate design tools and
client/user feedback to select
best solution
Some factors missed in
evaluating design solution
Compares the design solution
against the project objectives
and functional specifications,
providing qualitative
evaluation where appropriate
Comprehensive evaluation of
design solution, with welldefended recommendations for
future work or implementation
In addition to the indicators described above, two professors in the Faculty of Engineering and Applied
Science have administered a test of design process understanding at the beginning and end of their
courses for the past three years.8 APSC-100 was the context in first year, and an elective multidisciplinary project-based design course, APSC-381, was the context for the third year assessment. Since
students elect to take APSC-381 the results should not be considered representative across the third
year of all programs, but provide a point of comparison for first year students.
The test instructions were as follows:
Situation:
You are leading a team of engineers on a design project, and need to plan out how you would approach
a problem.
Instructions:
1) Select and circle one scenario from the box below.
Content for this was adapted from a published study: B. Frank, D. Strong, “Development of a Design Skill
Assessment Tool”, Canadian Engineering Education Association Conference 2010, Kingston, ON, available online
at: http://library.queensu.ca/ojs/index.php/PCEEA/article/view/3165/3103
8
49
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
2) Outline, on the back of this page, the specific steps you would take to ensure a successful design for
your selected scenario. Please ensure you:

Outline all special considerations and specific requirements associated with your selected
scenario.

Outline the design process you would follow to solve the problem, listing specific steps your team
would take. You do not need to create a specific design, sketch, etc.
The students selected one scenario from a list of three provided on each test. An “A” version, with 3 of
the scenarios, and a “B” version, with three different scenarios, were each given to half of the class as
the pre-assessment, and the versions switched for the post assessment. This way the average difficulty
of the two versions could be evaluated.
Student responses were evaluated against 6 commonly accepted stages of engineering design: problem
definition, preliminary design, conceptual design, detailed design, validation, and implementation. They
are also assessed for the overall process (iteration and appropriate order) and use of design tools.
The figures below show the distribution of “pre” and “post” subtotal scores on the for the groups of
interest. The largest gains occur in APSC381, as would be expected since it is an advanced 12-week
course on design tools and methodologies. APSC100 students also show significant gains in the two
deliveries. Very modest gains are apparent from the outreach group. Note that the distribution of
“post” scores for APSC-100 is higher than “pre” scores for APSC-381, possibly due to low knowledge
retention between first and third year, or limited design instruction between first year and winter
semester of third year.
The histograms for APSC-100 show students making gains in conceptual and preliminary design, which
include activities like idea generation, planning, comparing potential designs, and early prototype
construction. This reflects the emphasis in the course. Note that score changes corresponding to a major
tick on the x-axis are grouped with the lower interval.
The corresponding data for APSC-381 students show much larger gains are made in problem definition,
conceptual design, preliminary design, and detailed design. This reflects the nature of the course which
does not require students to implement their designs. An even more significant increase in score for use
of Tools can be seen students were noted to make reference to design tools in the post-course
assessment.
50
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
PRE
Figure 9: Change in score from in 2010 “pre” to
“post” for APSC100 students in the major
categories. Statistically significant changes at a
95% confidence level are indicated with an
asterisk.
POST
Figure 8: Distribution of scores in 2010 for pre (top)
and post (bottom) engineering activity. Boxes show
middle quartiles, dark line shows median score.
Figure 10: Change in score in 2010 from “pre” to
“post” for APSC381 students in the major categories.
Statistically significant changes at a 95% confidence
level are indicated with an asterisk.
51
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Scoring rubric used to assess design skill assessment scenarios.
52
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Teamwork
Plots of performance scores on teamwork indicators are shown below. Generally performance on these
was acceptable. However, for this year no descriptors were provided to graders on these indicators (as
shown below). Given the large percentage of students scored as “Exceeding expectations”, descriptors
will need to be provided to provide guidance to graders in future years.
Figure 11
Threshold
Target
1 - Not
Demonstrated
2 - Marginal
3 - Meets
Expectations
4 - Outstanding
3.06 - FY1: Recognizes a variety
of working and learning
preferences
Fail
Marginal
Good
Excellent
3.06 - FY2: Applies principles of
conflict management to resolve
team issues
Fail
Marginal
Good
Excellent
3.06 - FY3: Assumes
responsibility for own work and
participates equitably
Fail
Marginal
Good
Excellent
3.06 - FY7: Exercises initiative
and contributes to team goalsetting
Fail
Marginal
Good
Excellent
53
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Communications
In addition to the faculty wide indictors of the communications attribute, some course specific indicators
for communications were developed and assessed as well. Generally students were observed to do well
on formal oral presentation expectations but poorer on expectations for written work. Throughout the
year approximately 15% of the students struggled with summarizing and paraphrasing written work
accurately. At the end of the semester approximately 15% of the student teams were deemed to be
below target on written communications, as measured by their final report.
450
400
350
300
250
1
200
2
150
3
100
4
50
0
FEAS - 3.07-FY3 FEAS - 3.07-FY3 FEAS - 3.07-FY3 FEAS - 3.07-FY3
(PropReportPM) (PropReportFS) (PropReportAA) (T2Scope&Info)
APSC100
APSC100
APSC100
APSC100
Figure 12
3.07 - FY3: Summarizes and
paraphrases written work
accurately with appropriate
citations
Threshold
1 - Not Demonstrated
2 - Marginal
Insufficient content
to assess summary of Records information from
work. Summary
few resources. Misses
misinterprets
significant points of view.
researched material.
Target
3 - Meets Expectations
Summarizes and paraphrases
written work accurately.
4 - Outstanding
Synthesizes main ideas to
construct new concepts.
Summarizes the leading thoughts
in the field and gives a broader
picture of the problem.
54
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Figure 13
1 - Not
Demonstrate
d
Threshold
Target
2 - Marginal
3 - Meets
Expectations
4 - Outstanding
Report achieves
goal using formal
tone, properly
formatted,
concisely written,
appropriate use of
figures, few
spelling/grammar
errors
Professional tone,
convincing
argument,
authoritative, skillful
transitions
Concise prose, few
extraneous words.
Little irrelevant
information
Concise, meaningful
text, with no
irrelevant
information.
Appropriate use of
varied sentence
structures.
Final Report
Written
Communication
Poorly
constructed
report
Some
organization
problems, minor
formatting
problems,
redundancy,
spelling/grammar
errors
Team Assignment 2
Communication:
Conciseness
Wordy
expressions,
including
"This is/There
is/It is"
structures.
Many
redundancies
or gaps
Some irrelevant
information,
wordy phrases,
redundancies,
and/or gaps
55
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Figure 14
Threshold
Target
3 - Meets
Expectations
1 - Not
Demonstrated
2 - Marginal
Proposal Report Coherence &
Format: Sequence,
transitions,
formatting
Poorly prganized;
rambling, lacks
unity;
Inconsistent
writing/formattin
g; lacks
transitions; many
gaps or
redundancies.
Handwritten
material
Organization
sometimes
unclear; some
unsuitable
sections, minor
gaps or
redundancies,
minor
formatting
problems; some
wordy
expressions
"This is/There
is/It is"
Organized,
appropriate
sections,
uniformly and
correctly
formatted; little
irrelevant
information.
Focused, logically
organized, skillful
and varied
transitions.
Professionally
formateed. No
irrelevant
information
Proposal Report Correctness
Numerous
grammar errors;
disorganized
sentence
structure;
inappropriate
units or
significant digits;
acronyms not
defined
Some grammar/
punctuation
errors; spelling;
incomplete
sentences.
Relatively free of
grammar/
punctuation
errors; units
specified;
acronyms and
symbols defined
No typos or
grammatical errors;
attractively and
professionally
formatted.
4 - Outstanding
56
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Proposal Report Graphical
Communications
Figures and tables
not related to
text. Improper
referencing of
figures. List of
tables/figures
missing.
Some figures
and tables not
discussed in
text.
Figure/table
captions
missing.
Incomplete/
inconsistent list
of tables/
figures.
Figures and tables
referred to in text,
captioned.
Appropriate lists
of figures/tables.
Figures and tables
professionally
formatted,
integrated into text,
complementing text
Proposal Report Purpose & Style
Unclear purpose.
Tone and style
inappropriate.
Audience needs
not met.
Some aspects of
presentation
don’t achieve
purpose.
Clear purpose is
met. Formal tone
and style
appropriate to
audience
Professional tone
and style.
Authoritative and
convincing
450
400
350
300
250
1
200
150
2
100
3
50
4
0
Prop Pres Oral Final Pres Oral Final Pres Oral Prop Pres Oral
Comm: Content Comm: Content Comm: Content Comm: Content
(FS)
(FS)
(PM)
(PM)
APSC 100
APSC 100
APSC 100
APSC 100
Figure 15
57
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
1 - Not
Demonstrated
Oral
Communication Content
Oral
Communication Organization:
Sequence;
transitions.
Methods to
emphasize content
(illustrations,
repetition)
Oral
Communication Purpose & Style
Oral
Communication Slides: Mechanics,
acronyms, units,
graphics
Oral
Communication Speakers: Volume,
speed, cadence,
English usage;
questions. Use of
aids
2 - Marginal
3 - Meets
Expectations
Appropriate
content covering
project
background,
scope, relevant
information,
design options,
and work
completed.
4 - Outstanding
Minimal
content that
does not
demonstrate
appropriate
progress on
project.
Generally
appropriate
content but
some
information
missing.
Insufficient
content; poorly
organized;
rambling; not
understandable
Organization
sometimes
unclear; some
unsuitable
sections;
disconnected.
Some useful
information
missing.
Organized;
appropriate
information;
generally smooth
transitions
between topics
Slides logical, easy
to follow; detailed,
impressive depth of
thought; excellent
transitions between
topics
Some aspects of
presentation
don’t achieve
purpose.
Clear purpose is
met. Formal tone
and style
appropriate to
audience
Professional tone
and style.
Authoritative and
convincing
Relatively free of
typos and
grammatical
errors. Slides
uniformly
formatted.
Acronyms and
symbols defined;
appropriate
graphics
Slides free of typos
and grammatical
errors; slides
uniformly formatted
and professional;
material and
graphics support
main ideas
Understandable,
appropriate
volume, pitch and
speed; questions
generally
answered
correctly; aids
used properly
Professional
variation in volume,
pitch, speed; very
knowledgeable and
clear when handling
questions; aids used
very well
Unclear
purpose. Tone
and style
inappropriate.
Audience needs
not met.
Numerous
typos and
grammatical
errors;
unprofessional
and inconsistent
formatting;
figures and
tables unclear
or unrelated
Hard to
understand/hea
r the speaker;
speaking far too
fast/slow;
questions not
satisfactorily
answered;
reading from
notes
Some typos;
some
inconsistency in
formatting;
some acronyms
and symbols not
defined;
inappropriate
use of
figures/tables
Generally
understood with
some volume,
pitch and speed
issues; some
questions
answered
incorrectly;
reading off of
slides
Excellent content
and evidence of
detailed thought
and analysis.
58
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
450
400
350
300
250
200
150
100
50
0
1
2
3
4
Prop Pres Oral Prop Pres Oral Final Pres Oral Final Pres Oral
Comm:
Comm:
Comm:
Comm:
Organization (FS) Organization Organization (FS) Organization
(PM)
(PM)
APSC 100
APSC 100
APSC 100
APSC 100
Figure 16
400
350
300
250
200
1
150
2
100
3
50
4
0
Prop Pres Oral Prop Pres Oral Final Pres Oral Final Pres Oral
Comm: Purpose Comm: Purpose Comm: Purpose Comm: Purpose
& Style (PM)
& Style (FS)
& Style (FS)
& Style (PM)
APSC 100
APSC 100
APSC 100
APSC 100
Figure 17
59
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
400
350
300
250
200
1
150
2
100
3
50
4
0
Prop Pres Oral Prop Pres Oral Final Pres Oral Final Pres Oral
Comm: Slides Comm: Slides (FS) Comm: Slides (FS) Comm: Slides
(PM)
(PM)
APSC 100
APSC 100
APSC 100
APSC 100
Figure 18
450
400
350
300
250
1
200
150
2
100
3
50
4
0
Prop Pres Oral Final Pres Oral Prop Pres Oral Final Pres Oral
Comm: Speakers Comm: Speakers Comm: Speakers Comm: Speakers
(FS)
(FS)
(PM)
(PM)
APSC 100
APSC 100
APSC 100
APSC 100
Figure 19
60
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Professionalism
Individual professional behaviour, punctuality, responsibility, and communication etiquette was
assessed on the team design projects at the middle and end of the projects. Additionally, a coursespecific indicator was assessed that required students to deal with an ethical dilemma using PEO’s code
of ethics (this indicator overlaps with the Ethics and Equity attribute). The student scores are shown
below. Generally graders felt that the students showed appropriate professional behaviour on teams,
but many students’ performed below expectation when it came time to apply a code of ethics to an
ethical dilemma.
Figure 20
3.08 - FY2: Demonstrates
punctuality, responsibility and
appropriate communication
etiquette
Ethical
dilemma
(PSM4)
1 - Not Demonstrated
Threshold
2 - Marginal
Target
3 - Meets Expectations
4 - Outstanding
Fail
Marginal
Good
Excellent
1
(not demonstrated)
No process to deal
with dilemma.
2
(marginal)
Process to deal with
dilemma based on
personal beliefs only.
3
(meets expectations)
Process to deal with
dilemma based on
personal beliefs and
applicable codes of
ethics.
4
(outstanding)
Process to deal with
dilemma based on
personal beliefs and
applicable codes of ethics
addresses obligations to
multiple parties.
61
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Impact of Engineering and Ethics and Equity
As shown in the ethical dilemma in the previous section, and in the plot below, a significant number of
students were below expectation in applying code of ethics, and making decisions based on their
potential impact on society.
Figure 21
1 - Not Demonstrated
3.09 - FY4: Incorporates
sustainability considerations in No consideration of
decision making (societal and
these factors.
ecological)
3.10 - FY3: Describes ethical
issues and how they affect the No or trivial
individual, the company and the recommendations
public
Threshold
2 - Marginal
Target
3 - Meets Expectations
Incorporated appropriate
Factors mentioned but no
social, environmental, and
clear evidence of impact on
financial factors in decision
decision making.
making
Some important
recommendations and
implications missed
Makes supported
recommendations to prevent
future failures; describes
implications to stakeholders
4 - Outstanding
Well-reasoned analysis of these
factors, with risks mitigated
where possible
Meets expectations PLUS:
Information, model, and
conclusions used to make
recommendations supported by
authoritative information
62
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Economics and Project Management
Generally first year students were observed to plan and manage time and money effectively at the end
of the semester-long design project (3.11-FY1). However, many teams struggled to create an
appropriate project scope and project plan early in the project.
Figure 22
1 - Not Demonstrated
3.11 - FY1: Plans and efficiently
manages time and money
Threshold
2 - Marginal
No useful timeline or
Poor timeline or budget;
budget described;
infrequent meetings; minor
poorly managed
safety problems
project; safety issues
3.11 - FY2: Establishes
appropriate project scope, after Scope of project is not
consultation with client, based defined.
on available resources.
Project scope not
appropriate for available
resources and project
objectives.
3.11 - FY3: Plan maps out project Plan does not refer or Plan has a general outline
with clear milestones and
map out a timeline
of milestones with some
delegation.
for the project
reference to duration
Target
3 - Meets Expectations
Plans and efficiently manages
time and money; team
effectively used meetings;
safety considerations are clear
4 - Outstanding
Efficient, excellent project plan
presented; detailed budget;
potential risks foreseen and
mitigated
Project scope definition clearly
Establishes appropriate project defended by economic analysis,
scope using cost/benefits
available time and skills, and
linked to project timeline.
Plan maps out project with
clear milestones and
delegation appropriate to
project stage.
Detailed layout (e.g. Gantt chart)
with clear plan including
consideration for item
dependencies as well as room for
readjustment and remedial action
63
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Lifelong learning
Indicators for lifelong learning at the first year level focus primarily on information gathering and
evaluation. The most significant issue identified here is critical evaluation of information sources, as has
been observed in previous years.
450
400
350
300
250
1
200
2
150
3
100
4
50
0
FEAS - 3.12-FY1 FEAS - 3.12-FY2 FEAS - 3.12-FY5 FEAS - 3.12-FY6
(T2Scope&Info) (T1ProjectDef) (T1ProjectDef) (T2Scope&Info)
APSC100
APSC100
APSC100
APSC100
Figure 23
1 - Not Demonstrated
3.12 - FY1: Uses information
effectively, ethically, and legally
to accomplish a specific
purpose, including clear
attribution of Information
sources.
3.12 - FY2: Identifies a specific
learning need or knowledge
gap.
Target
3 - Meets Expectations
4 - Outstanding
No way to check
validity of
information.
Inconsistent use of citation
Clear attribution of all sources,
guidelines; some
Outstanding range of sources,
including graphics, using a
information sources
cited properly.
single citation style
ambiguous
No description of
learning need.
Some learning needs
identified but some gaps
remain.
Identifies learning needs or
knowledge gap based on
project requirements
Identifies learning needs
encompassing all phases of
project.
Some identified needs not
appropriately met by
identified information
sources.
Identifies appropriate
technical literature and other
information sources to meet a
need
Identifies multiple authoritative
sources from diverse
perspectives. All source selection
is justified.
Presents information that is
not relevant. Accepts all
information found. Does
not check for timeliness.
Describes professional-grade
Presents information that has
sources and reputable experts in
clearly been critically
technical, regulatory, and social
evaluated based on authority,
aspects, as appropriate.
currency and objectivity
3.12 - FY5: Identifies appropriate
technical literature and other
No useful information
information sources to meet a sources identified.
need
3.12 - FY6: Critically evaluates
the procured information for
authority, currency, and
objectivity.
Threshold
2 - Marginal
Presents information
that is not relevant.
Accepts all
information found.
Does not check for
timeliness.
64
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Overall analysis
The assessment of Knowledge Base in first year focused on introductory calculus, APSC-171. Based on
the data, the focus in the next year will be on helping students develop the ability to create a
mathematical model of a real-world problem, and on integration activities.
At the first year level, the majority of the attributes were assessed in APSC-100. In this course there
were several indicators on which students were scored low multiple times. The list of indicators on
which more than 20% of the students fell below the target expectation (“meets expectations”, or 3 out
of 4) were as follows (the number of times 20% or more of the students fell below the target is shown in
parentheses):







3.02-FY1: Identify problem (twice)
3.02-FY4: Evaluate solution (three times)
3.07-FY3: Summarizes accurately (only for proposal, this significantly improved by the end of the
first year)
3.10-FY3: Describes ethical issues (twice)
3.11-FY2: Project scope and 3.11-FY3: Project plan – these were weak on two deliverables in the
middle of the winter semester
3.12-FY5: Identifies appropriate literature
3.12-FY6: Evaluates information
Some of these were reassessed in other ways at the end of the first year, and the overall class
performance was much stronger at the end. At the end of the first year students were also evaluated on
overall written communication, and 16% of the teams were deemed below target.
Student performance on indicators in APSC-100 was examined over the year to examine how students
were adapting to expectations. The figure below shows the mean score on activities, and the percentage
of the students below target, as a function of the date in which the activity occurred. Early activities
more likely to be below threshold, and by the end of the year the percentage of students below
threshold was much lower than at the beginning. The trend lines suggest that students are adapting to
overall expectations within the course.
65
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
45
4.000
40
3.800
% Below target
3.600
Mean
3.400
30
3.200
25
3.000
20
2.800
15
2.600
10
2.400
5
2010-08
Mean score
Percent below target
35
2.200
2010-09
2010-11
2011-01
2011-02
2.000
2011-04
Approximate deliverable date
Figure 24 - Percent of students below target (red) and mean score on indicators scored out of four (blue) by date
The performance of students across all the indicators was also analyzed. The figure below shows the
number of students scoring below the target level of expectation on the indicators. Note that these
include faculty –wide indicators and course-specific indicators. Approximately half of the students did
not fall below target level on any indicator (this includes group work). Approximately 80% of the
students fell below threshold on 2 or fewer indicators.
66
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Figure 25
When the comparison is restricted only to faculty-wide indicators, the result is shown below. About half
of the students do not fall below the target expectation for any indicator, and 70% fall below for 2 or
fewer indicators. Approximately 80% do not fall below threshold level on any indicator, and 96% fall
below threshold for two or fewer indicators.
67
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
700
Number of students
600
584
500
Below target
400
300
Below threshold
293
200
96 108
61 48 55
100
56
54
37
16
6
0
0 13 0
30
00
00
00
00
00
0
0
1
2
3
4
5
6-10 11-1516-2021-2526-3031-3536-4041-50
Number of indicators
Figure 26
Plans for program improvement
At the first year level, specific plans for improvement based on the data include:




Grader calibration: In analyzing the data it was also observed that the variation between
graders was larger than desirable. In the previous year students were assigned to sections
randomly, and all students are taught by same instructor, but graded by section. An analysis
of average scores assigned by graders show statistically significant differences at the .05
level between graders scores on deliverables. As a result greater effort will go into grader
training and calibration
Problem analysis: greater focus will be placed on making an effective argument, and
comparing the outcome of a solution to the originally defined problem definition.
Design: the focus on safety and risk assessment is going to be expanded in the coming year
(overlaps with attributes Professionalism and Impact on Society), including focus on
Occupational Health and Safety.
Communications: Student feedback and performance indicate that communication skill
development needs to be revised in first year. Additional effort will be provided to support
students with weak communication skills at the beginning of first year. If an individual or
68
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program




group falls below some level of expectation in communications they will be required to
meet with the TA for detailed feedback, and pointed at resources that they can use to
improve the document before resubmitting.
Economics and project management: additional emphasis is being placed on problem
definition, project scoping and planning
Lifelong learning: additional emphasis is being placed on evaluating information and making
an effective argument
On some indicators a large number of students are above expectations, suggesting that the
expectations may be too low
Based on survey results, student view ethics and equity, professionalism, and impact of
engineering as receiving lower priority within program. This is being emphasized in the
engineering design and practice stream being introduced in all programs
Faculty-wide middle year assessment: Communications
These two courses taken by most students in second year provide basic instruction in Engineering
Communications. Full details of their delivery are available in the supplemental materials filed with the
Course Information Sheets. The course activities were loosely tied to technical and design activities in
other courses, varying with program. The rollout of our new common design course APSC 200 in the
2011-2012 academic year will allow tight integration of the communications elements of the new APSC
293 with student design projects in second year across all programs.
The focus of both courses was on student communication activities, practice, and revision in workshops.
There were five modules in each of the 12 AU courses, each equally weighted. Each course included an
individual presentation module, a group presentation module, and three written modules. They were:







Job Application, including resume and cover letter. (291 JA)
One Minute Engineer, a brief extemporaneous oral presentation, followed by written reflection
on video of the performance and means for improvement. (291 OME)
Briefing Note, a three page written analysis of the potential for district heating at Queen’s. (291
BN)
Incident Report, a written report of a fictitious machine shop accident, using mandated forms
and a follow-up memo. (291 IR)
Group Presentation, with supporting slides, on a technical topic chosen by the students. (291
GP)
Executive Summary, from review of a report prepared by 4th year students. (292 ES)
Team Oral Communication, in a design group meeting with the aid of a whiteboard, followed by
written reflection on performance and means for improvement. (292 TOC)
69
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program

Procedures and Systems, detailing a sequence in a process and/or a set of instructions,
complete with block diagram. (292 PS)
 Discussion and Conclusions, communications and reasoning on a technical topic chosen by the
students. (292 DC)
 Group Presentation, with supporting slides, on a technical topic chosen by the students. (292
GP)
To pass, students were required to achieve a satisfactory outcome in all modules. Revision cycles in
workshops and resubmission allowed all students the opportunity to improve their work. (Incorporation
of specific, marked revision cycles represents an improvement over previous offerings with a single final
submission of each deliverable.) Specific communications outcomes from the CDIO Syllabus were
evaluated for each module, as well as a separate outcome for overall performance on the module. The
overall performance was not a simple sum of the other outcomes, thus a submission could be
unsatisfactory and require resubmission even if 3 out of 4 subsidiary outcomes were satisfied. CDIO
Syllabus topics were used in part because of the instructor’s familiarity with them, and in part because
the faculty outcome list had not been finalized when the course design was being completed. The
outcomes used will be harmonized as part of the roll-out of the faculty-wide integrated design and
communications courses APSC 200/293 in Fall 2011. All of the outcomes used are contributors to
meeting Graduate Attribute 7, Communications.
Data were accumulated for each outcome during the marking process and stored in the Moodle course
management system. The charts that follow were completed based on Moodle records at the end of the
academic year. They do not reflect the subsequent resubmission and success of a small number of
students who completed modules after the end of classes. In addition, some subsidiary outcomes were
not recorded when students completed revisions in workshops, resulting in lower numbers of
observations. The data for all students in both courses was downloaded, aligned to provide a complete
record for each student, and then anonymized for privacy. The resulting frequencies were accumulated
over all students and plotted for each outcome in each submission of each module. The overall data set
includes 483 students, the vast majority of whom participated in both courses.
Each of the outcomes were ranked on an Engineering Practice scale of:
 Resubmit: The work is unsatisfactory and must be revised.
 File Only: The work is sufficient to provide a record for the file.
 Internal Circulation: I would show this to my boss.
 Almost Client Ready: Only a few minor tweaks are required.
 Client Ready: The work is ready for submission to an outside client with no revisions.
All of the outcomes are satisfactory except Resubmit. File Only is below expectations, but sufficient to
pass the course. More details accompany the Course Information Sheets.
Students generally performed well on the communications tasks set for them. There is evidence that
they were able to improve their outcomes through a cycle of revisions. Any improvements over the
course of the year are masked by other factors, including rising expectations of academic assistants over
the course of the year, student workloads, course administration glitches, etc. More telling comparisons
70
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
are available between second year and final year outcomes displayed in communications elements of
capstone projects.
Over the course it became evident that students were not expecting the graders to apply the strong
standards we advertised for the first assignment, and the vast majority of students were required to
resubmit their job applications, mostly on matters related to format, spelling, punctuation and
grammar. Most students can meet targets of either Internal Circulation or Almost Client Ready on the
first try. Student performance on each module after their revision process in workshops and
resubmission of their work shows that more than 85% of the students met expectations with scores of
Internal Circulation or above and virtually all students were able to meet a minimum threshold standard
of File Only after multiple revisions.
Several areas can be identified where students show higher frequencies of the marginal File Only
outcome were also identifiable from anecdotal experience. These items should receive more attention
in the APSC 200/293 offering in 2011/12:


Students had more difficulty identifying key elements for a summary and expressing them
clearly and concisely.
More students performed below expectations in informal small group discussions than in a
formal presentation, and they were less able to identify their problems and opportunities to
improve on them.
The chart below shows APSC 291/292 outcomes related to the mechanics of written communication.
The outcomes on the left stand out because the Job Application module required students to resubmit
at least twice to eliminate any errors in format, punctuation, spelling or grammar. The vast majority of
students meet expectations on all assignments and can exceed expectations when required.
71
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
72
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
This chart shows APSC 291/292 outcomes related to critical thinking in communications.
73
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
This chart shows APSC 291/292 outcomes for oral presentations, both extemporaneous, individual presentations and more formal group presentations with
slides.
74
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Faculty-wide Middle year assessment: Economics
The table below shows the percentage of the course assessment on the five indicators. These include all
assignments (5 individual 2, team), and questions on the midterm and the final exam.
Code
3.11-MY1
3.11-MY2
3.11-MY3
3.11-MY4
3.11-MY5
Indicator
Gather appropriate information, categorize it, and determine the economic
attractiveness of an engineering project
Measure the financial impact of risks that are associated with the
engineering project and consider the risk and return relationship as a
component of determining economic attractiveness
Plans a project from scoping to completion, including scheduling
Plans for the management of risks and changes required for project
Describes new enterprise formation, business planning and the economic
attractiveness of the enterprise.
% of course
assessment
41
3
15
8
34
The description under Economics and project management speaks specifically to indicators 1, 2, 3 and 4
while indicator 5 was included as important related material.
Student performance on activities was analyzed after the course. Students were scored on a numeric
scale appropriate to the deliverable, rather than using a rubric as in other courses. Overall student
performance on the five indicators measured on questions given on the midterm and final exams are
shown below for comparison.
Code
MY1
Indicator
Economic
attractiveness
Indicator
Gather appropriate information, categorize it, and
determine the economic attractiveness of an engineering
project
MY3
Project Plan
Plan the project from scoping to completion, including
scheduling
MY4
MY5
Manage Risks
Describe
Enterprise
plan for the management of risks and changes required
Describe new enterprise formulation, business planning
and enterprise economic attractiveness
AVG High Low
59 100
0
63 100
5
79 100
35
95 100
0
70 100
0
82 100
45
74 100
0
35 100
0
69 100
15
78 100
30
MY5, describing enterprise formulation and economic attractiveness, had the lowest average score.
75
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Observations
Slightly less emphasis can be placed on MY1 (economic attractiveness) and a lot less on MY5 (new
enterprise formation). While MY5 is reduced overall the Capital asset Pricing Model (CAPM) reflecting
cost of equity for the Enterprise must be included. Slightly more emphasis will be placed on developing
and assessing MY2,MY 3, and MY4 (risk assessment, project scoping, and risk management).
Approach for 2011-12:
One more indicator will be added to the course activities:
Describe a project's sustainability and broader contribution and impact on the enterprise,
environment and society.
Overall the course learning outcomes will be the six economic indicators. The course delivery and
assessment will be adjusted to provide a balance in assessment.
Code
3.11-MY1
3.11-MY2
3.11-MY3
3.11-MY4
3.11-MY5
3.11-MY6
Indicator
Gather appropriate information, categorize it, and determine the economic
attractiveness of an engineering project
Measure the financial impact of risks that are associated with the
engineering project and consider the risk and return relationship as a
component of determining economic attractiveness
Plans a project from scoping to completion, including scheduling
Plans for the management of risks and changes required for project
Describes new enterprise formation, business planning and the economic
attractiveness of the enterprise.
Describe a project's sustainability and broader contribution and impact on
the enterprise, environment and society.
% of course
assessment
35
5
20
10
20
10
76
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Student input on Graduate attributes
The faculty conducted a survey across all programs to understand student perceptions of the graduate
attribute requirements, and to solicit their input into how the program can be developed to help them
develop the required attributes. This was done by two methods:
1. Web-based surveys of first year and graduating year students. Survey questions are shown in
Appendix 3.1B.
2. Interviews with members of the undergraduate engineering student society.
This study was approved by the Queen’s University General Research Ethics Board.
First year survey
Of the approximately 650 students in the first year program 188 students (29%) responded to the
survey, and 129 of those (20%) completed the entire survey.
Students were asked to report the degree to which they felt they already possessed attributes upon
entry to the program. The top five graduate attributes where students reported a rating of 2 or 3 (yes or
to a great degree) out of three is as follows:
Individual and Team Work
Communication Skills
Professionalism
Problem Analysis
Investigation
89%
78%
69%
61%
61%
Students were then asked to report on the degree to which first year programming supported
development of the attributes. The top five graduate attributes where students reported a rating of 2 or
3 (yes or to a great degree) out of three is as follows:
Individual and Team Work*
Knowledge base in engineering
Problem analysis*
Professionalism*
Investigation*
Design
Impact of Engineering on Society
95%
94%
94%
86%
83%
81%
81%
*Identified as a strength coming in to the program
77
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Knowledge base and problem solving were most frequently listed as the top priority in the program as
shown in the table below. This is not expected in an engineering program, particularly in first year when
the majority of the courses focus on natural sciences and mathematics. Interestingly, teamwork,
communications, and professionalism were also frequently listed.
Attribute
Problem Solving*
Individual & Team Work*
Knowledge Base
Communication*
Professionalism*
# of students who incl it in
top 5
120
97
95
76
47
# of students who selected it
as #1 priority
34
17
51
6
4
The three listed least frequently as being developed by first year programming were:
Ethics and Equity
Economics and Project Management
Lifelong Learning
64%
70%
73%
It is interesting that even the attributes with the lowest scores were perceived by approximately twothirds of the first year class as being developed in the first year of the program.
Attribute
Lifelong learning
Economics and Project
Management
Ethics and Equity
Use of Eng Tools
Impact on Society
# who included it in bottom 3
80
68
45
38
37
# who ranked it lowest
priority
38
25
1
15
6
78
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Fourth year surveys
The graduating year survey was sent to approximately 600 final year students. There were a total of 156
responses, and 102 completed the entire survey.
Students were asked which of the attributes they felt were difficult to demonstrate. Responses, along
with some general comments, are shown in the table below. There were 73 responses to this question.
Attribute
Ethics and Equity
# of students
25
Professionalism
Impact on Society
Communication
13
12
11
Design
9
Lifelong Learning
Economics & Project
Mgmt
8
8
Rationale
Next to no content since 1st year on these issues; student
perception of unethical behaviour in program (plagiarism
etc)
Lack of opportunity in the program (12)
Many assignments are team based so difficult for all
students to demonstrate competence; feedback on
written work takes too long to be able to improve for
next assignment; not always integrated
Design is integral, should be focal point of more courses;
focus on team work means that individual contributions
are difficult to recognize/demonstrate
More of an attitude than a skill
Lack of opportunity in the program; only touched on in
APSC 100 and APSC 221
Four of the students pointed out that, aside from knowledge base for engineering, most attributes are
difficult to demonstrate through traditional exams. There was a general perception that ethics and
equity, professionalism, and impact on society weren’t significant in upper years of programs. This has
was identified as an area of improvement several years ago, and will be discussed under program
development.
Students were asked to what degree did the engineering program support the development of
attributes. The table below shows the percentage of students who responded affirmatively (2) or to a
high degree (3) out of 3. The attribute for which the most students felt the program supported their
development to a high degree was individual and team work, and the lowest was ethics and equity.
Attribute
Knowledge base
Problem analysis
Ind & Team work
Investigation
Design
Use of eng tools
Communication
2+3
98.41
95.25
93.65
91.27
88.10
84.93
80.95
3 (to a high degree)
63.49
69.05
74.60
42.86
50.00
38.10
43.65
79
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Lifelong learning
Professionalism
Economics & pro mgmt
Impact on society
Ethics and equity
79.36
79.20
75.20
68.26
60.00
46.03
36.80
20.00
27.78
17.60
The attributes ranked by students as being given the highest priority most often were:
 Knowledge base for engineering (53)
 Problem analysis (31)
Those most often included in the top 5 were:





Problem analysis (105)
Knowledge base (92)
Individual & team work (80)
Design (79)
Communication (45)
Those ranked by students as being given the lowest priority the most often were:
 Ethics and equity (29)
 Impact on society (20)
Those most often included in the bottom five were:
 Ethics and equity (73)
 Impact on society (51)
 Professionalism (46)
 Lifelong learning (46)
With the exception of professionalism, the same attributes were ranked lowest by first year students.
80
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Individual interviews
Toward the end of the academic year students were asked to participate in focus groups and interviews.
At the time students were in exams and so it was difficult to get significant participation.
In the end two undergraduate engineering students were interviewed, both of whom were members of
the undergraduate engineering society executive elected by their peers. Generally these students are
better informed and more engaged in faculty matters than the average student.
One student had just finished second year, one just finished third year.
The most significant comments and recommendations from the students were:







Communicate graduate attributes and draw attention back to them. E.g. For example by
pointing out that assignments develop and demonstrate particular attributes
Use writing activities to support under-developed attributes in program (impact, ethics, etc.)
Make communications an important part of multiple courses, not only communications-only
courses
Don’t understand “lifelong learning” attribute
Professionalism and ethics and equity should be focused on in upper years
Portfolios would not be well received if they were an add-on to what’s already in the curriculum
Movement toward grad attributes is good. Need better communication about them and stress
that they are important elements of what it means to be an engineer
81
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Program development informed by data
Faculty wide development
Improvements in specific faculty-wide course were addressed in the previous sections on first and
middle year courses. This section will address a proposal for a sequence of four courses on Engineering
Design and Practice that has been in development for two years. This course will target 11 of the 12
graduate attributes; only knowledge base is not specifically targeted, as it is the primary attribute
developed in most engineering courses. As identified in the curriculum mapping exercise and student
surveys, development of these professional skills needs to be enhanced. The sequence will build upon
existing courses, including APSC-100 in first year, capstone design courses in the graduating year, and
some existing design and professional practice courses.
The first new course introduction happens in the 2011-2012 with a new pair of faculty-wide second year
courses, APSC-200 and 293. These will be taught by the same personnel, and are separated into two
courses for the purpose of record-keeping. APSC-200 focuses primarily on design and engineering
practice, and APSC-293 focuses on written and oral communications. The projects in the design course
will serve as the context for the reports and presentations used in the communications course. This will
provide multiple revision cycles to help students develop an effective writing process. The design
activities are also supported by materials from APSC 221/321 on engineering economics and business
practices.
In this project based design course, students will participate constructively on teams to create solutions
to open-ended problems, using standard design methods and tools. The course opens with a discipline
specific project activity in the first week to engage the students in an active learning opportunity.
Instruction will be provided primarily in the first 6 weeks of the semester focusing on problem scoping,
creativity and idea generation, decision making incorporating technical, economic, societal, and
environmental factors, safety, engineering codes and regulations, and engineering ethics. In that first six
weeks student groups complete a simple design/build project to provide a focus for tools they are
learning. The final 6 weeks of the course centers around a more advanced design project delivered by
each discipline. The communication course will include a small number of lectures, online modules, and
working tutorials supporting deliverables in the design course. This ensures the communications
expectations occur in an engineering context.
The course objectives include:



Employ and apply design processes and tools with emphasis on early stages (problem definition,
idea generation and decision making)
Promote creative processes in open ended problem solving
Design (involving analysis and/or simulation and/or prototyping) a solution for presented openended problem
82
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program


Excite students about engineering and the discipline they selected
Apply engineering principles and theories from other disciplinary courses to solve an openended problem
 Employ appropriate technical and non-technical factors in decision making, including some
combination of economic, environmental, safety, and social factors, and consideration of the
public interest
 Awareness of engineering as a regulated profession, including reference to some relevant
engineering regulations/codes/standards
 Awareness of professional/technical associations in engineering and discipline
 Awareness of occupational safety regulations
 Teaming skills
In future years third and fourth year courses will be introduced that build upon these objectives.
Section 3.1 Part B: Future plans for graduate attribute assessment
Programs should describe future plans for graduate attribute assessment in preparation for the next
accreditation cycle when criterion 3.1 will form a basis for accreditation decisions. Also describe the
assessment cycle if attributes will be assessed on a rolling basis.
In future years the program development process will be expanded to include more detailed data
gathering and analysis. Specific improvements include:










All 12 graduate attributes will be assessed in first year and the graduating year, and many will
also be assessed in middle years
Review and revise current indicators based on previous year use, including providing
information about the context in which each is evaluated
Program specific indicators will be developed for many of the attributes, including knowledge
base, engineering tools, and investigation
Faculty-wide and departmental committees will establish regular timelines for attribute
assessment, data analysis, and program improvement. These will include student members
Industrial feedback on the process and outcomes will be incorporated
Mapping of attributes in the existing program will be expanded to ensure that attributes are
being developed throughout the program
Faculty-wide design and professional practice courses will be used allow longitudinal assessment
of students throughout their program
Recently introduced and upcoming provincial and university policies for quality assurance will be
incorporated into the program development process
Policies will be established to deal with students who fall below threshold performance on
indicators. In some cases this will require students revise and resubmit material until threshold
performance is met
Instructors will describe relationship of course materials and deliverables to attributes
83
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program


Student portfolios will be introduced in selected courses to allow students to determine how
they meet expectations
Indicators will be assessed using multiple methods to assess reliability of data
84
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Appendix 3.1A: Levelled indicators
These indicators were developed in the 2009-2010 academic year. A selection of these were assessed in the 2010-2011 year based on timeliness
and the current state of the curriculum.
Sec
Attribute
First year
3.01-FY1
3.01-FY2
3.01
3.02
Knowledge
base
Problem
analysis
Middle years
Graduating year
Create mathematical descriptions or
expressions to model a real-world
problem
Select and describe appropriate tools
to solve mathematical problems that
arise from modeling a real-world
problem
3.01-FY3
Use solution to mathematical
problems to inform the real-world
problem that gave rise to it.
3.01-FY4
FCI score as indicator of basic physics
concepts
3.01-FY5
CCI score as indicator of basic
chemistry concepts
3.02-FY1
Identifies known and unknown
information, uncertainties, and
biases when presented a complex illstructured problem
3.02-GY1
3.02-FY2
Creates process for solving problem
including justified approximations
and assumptions
3.02-GY2
3.02-FY3
Selects and applies appropriate
quantitative model and analysis to
solve problems
3.02-GY3
Selects and applies appropriate
model and analysis to solve problems
3.02-FY4
Evaluates validity of results and
model for error, uncertainty
3.02-GY4
Evaluates validity of results and
model for error, uncertainty
Identifies problem, known and
unknown information, uncertainties,
and biases
Creates process for solving problem
including justified approximations
and assumptions
85
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Sec
3.03
Attribute
First year
3.03-FY1
Generates ideas and working
hypothesis
3.03-FY2
Designs investigations involving
information and data gathering,
analysis, and/or experimentation
Investigation
3.03-FY3
3.03-FY4
3.04
Design
3.04-FY1
3.04-FY2
Middle years
Graduating year
Synthesizes data and information to
reach conclusion
Appraises the validity of conclusion
relative to the degrees of error and
limitations of theory and
measurement
Adapts general design process to
design system, component, or
process to solve open-ended
complex problem.
Accurately identifies significance and
nature of a complex, open-ended
problem
3.04-FY3
3.04-GY1
3.04-GY2
3.04-GY3
Identifies customer and user needs
3.04-FY4
3.04-FY5
3.04-FY6
Gathers and uses information from
appropriate sources, including
applicable standards, patents,
regulations as appropriate.
Produces a variety of potential design
solutions suited to meet functional
specifications
3.04-GY4
3.04-GY5
Identify problem and constraints
including health and safety risks,
applicable standards, economic,
environmental, cultural and societal
considerations
Applies appropriate knowledge,
judgement, and design tools, in
creating and analyzing conceptual
design solutions to select best
concept
Creates and tests simulations,
models, and/or prototypes at various
points in design with complexity
appropriate to design stage
Assesses design performance based
on requirements, yield, reliability,
and/or safety as appropriate
Identifies possibilities for further
improvement and conducts design
review to evaluate performance of
the overall process.
Performs systematic evaluations of
the degree to which several design
concept options meet project criteria
86
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Sec
3.05
3.06
Attribute
Engineering
tools
Teamwork
First year
3.04-FY7
Compares the design solution against
the problem objective
3.05-FY1
Selects appropriate measurement
devices or techniques to accomplish
a task
3.05-GY1
Evaluates techniques, resources, and
tools to identify their limitations with
respect to needs
Applies appropriate techniques,
tools, and processes to accomplish a
task
Evaluates appropriateness of results
from several engineering techniques
and tools
Follows protocols when using
techniques, skills and tools
3.05-GY2
3.05-FY3
Demonstrates correct use of testing
apparatus, databases and models
3.05-GY3
3.06-FY1
Recognizes a variety of working and
learning preferences
3.06-GY1
Demonstrates capacity for initiative
and technical or team leadership
while respecting others' roles
3.06-FY2
Applies principles of conflict
management to resolve team issues
3.06-GY2
Evaluates team effectiveness and
plans for improvements
3.06-FY3
Assumes responsibility for own work
and participates equitably
Describes own temperament
3.07-GY1
Writes and revises documents using
appropriate discipline-specific
conventions
3.06-FY4
3.06-FY5
Critically analyzes results from a
temperament sorter, defending
opinion of how well results apply
3.06-FY6
Analyzes impact of own
temperament on group work
3.06-FY7
Exercises initiative and contributes to
team goal-setting
Writes using standard formats
3.07
Graduating year
3.05-FY2
3.07-FY1
Communications
Middle years
3.07MY1
Adapts format, content,
organization, and tone for
various audiences
3.07-FY2
Writes using standard grammar and
mechanics
3.07-GY2
Demonstrates accurate use of
technical vocabulary
3.07-FY3
Summarizes and paraphrases written
work accurately with appropriate
citations
3.07-GY3
Demonstrates confidence in formal
and informal oral communications
87
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Sec
Attribute
First year
3.09
3.10
Professionalism
Impact of
engineering
Ethics and
equity
Graduating year
3.07-FY4
Delivers clear and organized formal
presentation following established
guidelines
3.07-GY4
Uses appropriate referencing to cite
previous work
3.07-FY5
Creates effective figures, tables, and
drawings employing standard
conventions to compliment text.
3.07-GY5
Uses graphics to explain, interpret,
and assess information
3.08-FY1
3.08
Middle years
3.08-FY2
Describes role of protection of the
public and public interest in decision
making
Demonstrates punctuality,
responsibility and appropriate
communication etiquette
3.08-GY1
3.08-GY2
Demonstrates professional bearing
3.08-FY3
Participates actively in meetings,
helps to generate ideas
3.09-FY1
Describes relationship between
human activity and earth systems
3.09-GY1
3.09-FY2
Describes management techniques
for sustainable development
Describes relevant legal
requirements governing engineering
activities
3.09-GY2
3.09-FY3
3.09-FY4
Incorporates sustainability
considerations in decision making
(societal and ecological)
3.10-FY1
Demonstrates behaviour congruent
with academic integrity expectations
of university and faculty
3.10-FY2
Identifies and iterates items from the
professional codes of conduct
3.10-FY3
Describes ethical issues and how they
affect the individual, the company
and the public
Integrates standards, codes of
practice, and legal and regulatory
factors into decision-making
processes (as appropriate)
3.09-GY3
3.09-GYX
Considers economic, social, and
environmental factors and/or
impacts in decisions
Evaluates trade-offs among goals and
concepts
Explains the societal, enterprise,
and/or technical context of the
system
Evaluates technical, social, and
environmental trade-offs among
goals and concepts
88
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Sec
Attribute
First year
3.10-FY4
Plans and efficiently manages time
and money
3.11-FY2
Establishes appropriate project
scope, after consultation with client,
based on available resources.
Economics
3.11-FY3
Plan maps out project with clear
milestones and delegation.
3.11MY1
3.11MY2
3.11MY3
3.11MY4
3.11MY5
3.11MY6
3.12
Lifelong
learning
Graduating year
Describes consequences of deviating
from professional codes of conduct
and university code of conduct
3.11-FY1
3.11
Middle years
Gather appropriate
information, categorize it,
and determine the economic
attractiveness of an
engineering project
Measure the financial impact
of risks that are associated
with the engineering project
and consider the risk and
return relationship as a
component of determining
economic attractiveness
Plans a project from scoping
to completion, including
scheduling
Plans for the management of
risks and changes required
Describes new enterprise
formation, business planning
and the economic
attractiveness of the
enterprise.
Describe a project's
sustainability and broader
contribution and impact on
the enterprise, environment
and society
3.12-FY1
Uses information effectively,
ethically, and legally to accomplish a
specific purpose, including clear
attribution of Information sources.
3.12-GY1
3.12-FY2
Identifies a specific learning need or
knowledge gap.
3.12-GY2
Critically evaluates procured
information for authority, currency,
and objectivity.
Demonstrate skills of self-education
89
Mechanical Engineering, Queen’s University - Questionnaire for Evaluation of an Engineering Program
Sec
Attribute
First year
3.12-FY3
Makes accurate use of technical
literature and other information
sources.
3.12-FY4
Middle years
3.12-GYX
3.12-GYX
Describes own learning style
3.12-FY5
Identifies appropriate technical
literature and other information
sources to meet a need
3.12-FY6
Critically evaluates the procured
information for authority, currency,
and objectivity.
3.12-FY7
Describes connections between past
experience and future projects
3.12-FY8
Selects project planning strategies
based on past experience
3.12-GYX
Graduating year
Describes professional and academic
societies in the discipline and how
new knowledge enters discipline.
(Later year)
Identifies resources and professional
associations that address own
ongoing professional development
(Later year)
Recognizes the need to keep current
regarding new developments in field
(Later year)
90
Appendix 3.1B: Student surveys
Students in first and graduating years were asked to complete web-based surveys on their
perceptions of the graduate attributes. The questions from the fourth year survey are show
below; the first year survey was similar but with fewer questions.
Continuous Program Improvement to Support Engineering Graduate Attribute
Development: A pilot project
Survey Questions
Fourth-Year Students
The 12 Engineering Graduate Attributes identified and defined by the Canadian
Engineering Accreditation Board will be made available for students' reference
while completing the survey.
3.1.1 A knowledge base for engineering:
Demonstrated competence in university level mathematics, natural sciences,
engineering fundamentals, and specialized engineering knowledge appropriate to the
program.
3.1.2 Problem analysis:
An ability to use appropriate knowledge and skills to identify, formulate, analyze, and
solve complex engineering problems in order to reach substantiated conclusions.
3.1.3 Investigation:
An ability to conduct investigations of complex problems by methods that include
appropriate experiments, analysis and interpretation of data, and synthesis of
information in order to reach valid conclusions.
91
3.1.4 Design:
An ability to design solutions for complex, open-ended engineering problems and to
design systems, components or processes that meet specified needs with appropriate
attention to health and safety risks, applicable standards, economic, environmental,
cultural and societal considerations.
3.1.5 Use of engineering tools:
An ability to create, select, apply, adapt, and extend appropriate techniques, resources,
and modern engineering tools to a range of engineering activities, from simple to
complex, with an understanding of the associated limitations.
3.1.6 Individual and team work:
An ability to work effectively as a member and as a leader in teams, preferably in a multidisciplinary setting.
3.1.7 Communication skills:
An ability to communicate complex engineering concepts within the profession and with
society at large. Such abilities include reading, writing, speaking and listening, and the
ability to comprehend and write effective reports and design documentation, and to give
and effectively respond to clear instructions.
3.1.8 Professionalism:
An understanding of the roles and responsibilities of the professional engineer in society,
especially the primary role of protection of the public and the public interest.
3.1.9 Impact of engineering on society and the environment:
An ability to analyse social and environmental aspects of engineering activities. Such
abilities include an understanding of the interactions that engineering has with the
economic, social, health, safety, legal, and cultural aspects of society; the uncertainties
in the prediction of such interactions; and the concepts of sustainable design and
development and environmental stewardship.
92
3.1.10 Ethics and equity:
An ability to apply professional ethics, accountability, and equity.
3.1.11 Economics and project management:
An ability to appropriately incorporate economics and business practices including
project, risk and change management into the practice of engineering, and to
understand their limitations.
3.1.12 Life-long learning:
An ability to identify and to address their own educational needs in a changing world to
sufficiently maintain their competence and contribute to the advancement of knowledge.
1) With 0 being not at all and 3 being to a great degree, to what extent did the
Engineering Program support your development in:
a. A knowledge base for engineering
b. Problem analysis
c. Investigation
d. Design
e. Use of engineering tools
f. Individual and team work
g. Communication skills
h. Professionalism
i. Impact of engineering on society and the environment
j. Ethics and equity
k. Economics and project management
l. Lifelong learning
2) Please identify the 5 attributes that, in your opinion, are given the highest priority
in the Engineering Program. Use 1 to indicate the highest priority attribute; 2 to
indicate the next highest and 3 to indicate the next:
(all attributes listed; drop down boxes for ranking)
3) Which 3 graduate attributes, in your opinion, are given the least priority in the
Engineering Program? (Use 1 to indicate the least priority attribute; 2 to indicate the
next least and 3 to indicate the next.)
93
(all attributes listed; drop down boxes for ranking)
4) Which instructional/assessment strategies were most beneficial in allowing you to
develop each of the graduate attributes? Please select no more than 2 for each
attribute:
Lecture-based classes
Small group long-term projects
Field placements
Studio
Seminars
Small group, short-term projects
Problem-based learning
Labs
Problem sets
Other (please specify)
5) In your opinion and given current course structures and assessment
strategies, which attributes were the most difficult for you to demonstrate?
Why?
6) It is recognized that many opportunities outside the regular engineering
curriculum support the development of graduate attributes. Please indicate
your involvement in the activities listed below and the degree to which they
contribute to your development of graduate attributes:
N/A
1-3 hrs/wk
4-7 hrs/wk
more than 7
Student Government
Special Interest Clubs
Sports Clubs
Employment
Other
7) For purposes of program improvement we would like to be able to make
inferences about the ways program strengths and limitations affect acquisition of
engineering graduate attributes across levels of academic achievement. Please
94
select the grade range that most accurately reflects your overall academic
average in the Queen’s engineering program.
95
Appendix 3.1C: Definitions and Guidelines
Criterion 3.1
The institution must demonstrate that the graduates of a program possess the attributes under
the following headings. The attributes will be interpreted in the context of candidates at the
time of graduation. It is recognized that graduates will continue to build on the foundations that
their engineering education has provided.
3.1.1 A knowledge base for engineering: Demonstrated competence in university level
mathematics, natural sciences, engineering fundamentals, and specialized engineering
knowledge appropriate to the program.
3.1.2 Problem analysis: An ability to use appropriate knowledge and skills to identify, formulate,
analyze, and solve complex engineering problems in order to reach substantiated
conclusions.
3.1.3 Investigation: An ability to conduct investigations of complex problems by methods that
include appropriate experiments, analysis and interpretation of data, and synthesis of
information in order to reach valid conclusions.
3.1.4 Design: An ability to design solutions for complex, open-ended engineering problems and
to design systems, components or processes that meet specified needs with appropriate
attention to health and safety risks, applicable standards, and economic, environmental,
cultural and societal considerations.
3.1.5 Use of engineering tools: An ability to create, select, apply, adapt, and extend appropriate
techniques, resources, and modern engineering tools to a range of engineering activities,
from simple to complex, with an understanding of the associated limitations.
3.1.6 Individual and team work: An ability to work effectively as a member and leader in teams,
preferably in a multi-disciplinary setting.
3.1.7 Communication skills: An ability to communicate complex engineering concepts within the
profession and with society at large. Such ability includes reading, writing, speaking and
listening, and the ability to comprehend and write effective reports and design
documentation, and to give and effectively respond to clear instructions.
3.1.8 Professionalism: An understanding of the roles and responsibilities of the professional
engineer in society, especially the primary role of protection of the public and the public
interest.
96
3.1.9 Impact of engineering on society and the environment: An ability to analyze social and
environmental aspects of engineering activities. Such ability includes an understanding of
the interactions that engineering has with the economic, social, health, safety, legal, and
cultural aspects of society, the uncertainties in the prediction of such interactions; and the
concepts of sustainable design and development and environmental stewardship.
3.1.10 Ethics and equity: An ability to apply professional ethics, accountability, and equity.
3.1.11 Economics and project management: An ability to appropriately incorporate economics
and business practices including project, risk, and change management into the practice
of engineering and to understand their limitations.
3.1.12 Life-long learning: An ability to identify and to address their own educational needs in a
changing world in ways sufficient to maintain their competence and to allow them to
contribute to the advancement of knowledge.
97
Download