Creating an Educational Research Culture of Assessment for Programmatic Excellence

Page created by Ernest Salinas
 
CONTINUE READING
Creating an Educational Research Culture of Assessment for Programmatic Excellence
Creating an Educational
Research Culture of
Assessment for
Programmatic Excellence
APTA EDUCATIONAL LEADERSHIP CONFERENCE 2019

Benchmarks Taskforce

u   Peter Altenburger, Assoc Prof, Indiana Univ
u   Jacklyn Brechter, Assoc Prof, Chapman Univ
u   Gary Chleboun, Prof, Ohio Univ
u   Amy Heath, Assist Prof, Western Michigan Univ
u   Diane Jette, Prof Emeritus, MGH Institute of Health Professions
u   Denise Schilling, Prof, Western Univ of Health Sciences
u   Kimberly Topp, Prof Emeritus, Univ of California San Francisco
u   Barbara Tschoepe, Visiting Prof, Franklin Pierce Univ

                                                                      1
Presentation Goals

u   Illustrate the value of creating a culture of programmatic assessment
u   Integrate data from internal and external stakeholders
u   Differentiate between the value of programmatic and professional shared
    data assessment
u   Create a commitment to develop a professionally supported national
    program assessment instrument
u   Ensure that such an instrument is able to share data across multiple
    stakeholders to strengthen informed decisions and enhance ongoing
    improvement

Defining Assessment

u   Establishing clear, measureable outcomes for student learning
u   Ensuring that students have the necessary resources to achieve those
    outcomes
u   Systematically gathering, analyzing, interpreting, evidence to determine
    whether student outcomes are consistent with expected goals and
    objectives
u   Creating an ongoing process that uses assessment outcomes to inform
    decision-making about where and how to improve student learning

                                                                               Angelo, 1995

                                                                                              2
Defining Assessment

u   Students learn best when courses and activities are purposefully designed as coherent,
    integrated learning experiences, in which courses and out-of-class experiences build on
    and reinforce one another

    “When students can see connections between learning experiences, learning is deeper and more
    lasting”
                                         Graff, 2008

    "Successful learners build pathways, or make connections between what they are trying to learn
    and what they already know. .."
                                       Niemczyk, 2012

u   Collaborative curricular designs require collaborative assessment

Assessment
Culture

                                                                                                     3
A Culture of Assessment

u   What is your visceral response to this comment from a peer-faculty
    member?
     u   The accreditation self-study starts next week

u   What data can you gather quickly for the Dean's queries:
     u   How do our numbers of PhD/DSc faculty compare to those of other
         public/private or Masters/R1 institutions?
     u   What is our student indebtedness as compared to other peer institutions?

Assessment Culture Defined

u   Culture: "The set of shared attitudes, values, goals, and practices that
    characterizes an institution or organization" (Merriam-Webster Dictionary)
u   Assessment Culture: An environment that values reflection in action and
    reflection on action using a systematic method of collecting, analyzing
    and acting on data to inform decisions, implement improvements and
    aspire toward excellence. (Diane Jette)
u   Setting high expectations and holding each other accountable.
    (Jensen et al., 2017)
u   Fundamental processes for benchmarking or quality assurance are 1)
    identifying indicators of excellence, and 2) mechanism(s) for capturing
    performance (Heath et al., 2018)

                                                                                    4
...any field that deserves
to be described as a
profession must take on
the highest standard of
self-evaluation, both for
individual practitioners
and the entire
professional community.

Lee S. Shulman, PhD
President, Carnegie
Foundation for the
Advancement of Teaching

Levels of                    PROGRAMMATIC
                             INSTITUTIONAL
Assessment                   PROFESSIONAL

                                             5
Programmatic Assessment

u   Student learning is measured based on achievement of the established
    program outcomes --- faculty performance and resources are also
    included in the assessment

u   "Programmatic assessment is an approach in which routine information
    about the learner’s competence and progress is continually collected,
    analyzed and, where needed, complemented with purposively collected
    additional assessment information..." (Schuwirth et al., 2017)

Institutional Assessment

u   Institutions assess student outcomes as linked to the institutional mission,
    values and goals
u   As student learning is an institutional outcome, programmatic and
    institutional assessments are linked
u   Accreditation bodies are looking for longitudinal outcome data to
    demonstrate continuous quality improvement in programs
u   Challenges are found in identifying data elements that fit multiple
    professions and programs, and efficient methods to collect data
    longitudinally after graduation

                                                                                   6
Professional
Assessment

Professional Assessment

u   Define professional assessment – PROFESSIONAL EDUCATION
    u   Is only as good as the programmatic and institutional culture
    u   Collective commitment to each other

u   The value
    u   Organization literature
    u   Understanding learning processes and value of assessment to improve
    u   Make the cultural shift for data collection and assessment

                                                                              7
Professional Assessment

u   Shared Programmatic Assessment
     u   Definition
     u   Value
     u   Importance
u   Culture creates
     u   Expectation and need

Shared
Programmatic
Assessment

                                     8
Shared Programmatic Assessment –
Value for the Profession

u   Comparator professions overviews and examples
     u   AAMC Site www.aamc.org/data-reports
     u   AACP – Assessment and Accreditation Management System (AAMS)
          u www.aacp.org/research/assessment-accreditation-management-system-aams

u   What is possible for physical therapy given what other professions have
    accomplished?

Where is Shared Programmatic
Assessment in Physical Therapy?

u   The challenge is what to measure
     §    Breakdown the idea that we need to have a perfect measure
     §    Not everyone is the same

                                                                                    9
Where is Shared Programmatic
Assessment in Physical Therapy?

§   Challenges/Successes
§   Limitations
§   Pair-share on how to shift the culture
     §   uses
     §   benefits
     §   barriers

Solutions for Implementing Shared
Programmatic Assessment

u   How does the profession create a measure of quality education?
     u   Provide a discussion on the possible educational metrics
     u   Provide a discussion on the amount and types of data needed for critical
         assessment

                                                                                    10
Data

  u   Presenters will use data to review outcomes at the programmatic,
      comparative school, or collectively at the level of the profession.
  u   They will share educational trends across all institutions including
      student debt, program cost, faculty trends and other key
      administrative metrics.

Data Integration and Summary

u Examples           of data integration
  u Program          and Institution
  u Program and Profession
  u Program, Institution, and Profession

                                                                             11
Creating a                                                       AUDIENCE ACTIVITY
    Culture

    Example: From a Developing Program

           Blueprints                          Foundation                             Building
•   Planning                            •   Faculty Champion                •   Putting the plan to
                                                                                action…not just for CAPTE,
•   Master Assessment Plan (MAP)        •   Buy in from Leadership
                                                                                but for continuous quality
•   Details: who is responsible for          Program Director                   improvement
    collecting data, when, where to
    get the data, when to report the               Dean                     •   Continuously reviewing the
    data and to whom, the pre-                                                  plan and the timeline to
    established thresholds to trigger                                           refine the plan
    change
•   Plan and Timeline to Refine Plan

                                                                                                             12
Example: One University
 Various Strategies

 Resources like apartments are the same – but décor/utilization varies
 u   Culture of assessment is influenced by
      u   People
      u   Profession – expectations of the profession

The Building – Office of Institutional
Research & University Assessment Committee

 u   Mandated assessment services
 u   Alumni Survey
 u   Financial Aid Assessment
 u   Admissions data
 u   Research activity - grants & publication
 u   Student Outcomes
 u   Demographics – student & faculty
 u   Program completion data - retention
 u   Accreditation report/self-study

                                                                         13
The Apartments – individual programs

u   Varied Utilization
u   PT – representative attends CAPTE workshop and assists with data
    collection & analysis
u   PA – representative attends all faculty meetings, identifies assessment
    opportunity & performs analysis
u   Podiatry & Veterinary Medicine – representative administers, analyze &
    report on administrator effectiveness/evaluation
u   Osteopathic Medicine – representative collects overall curriculum &
    individual course effectiveness
u   Dental - Employment surveys

Questions for discussion/reflection

Many strategies and many resources available
u   Can we agree to a culture of assessment and a similar/standardized
    process?
u   Can we create a data bank of assessment options that allow programs to
    select the best option for them?
u   What is “your” culture of assessment?
u   How are you utilizing available resources?
u   How are you fostering a culture of assessment?

                                                                              14
Cultural Summary

    u    The movement toward comprehensive programmatic assessment has
         been around for over 25 years.

    u    Must Establish
          u   Value
          u   Importance

    u    Commitment from leadership is Key
          u   Programmatic leadership
          u   Campus Leadership
          u   Professional Leadership

    Sustained Professional Assessment

u   Striving for Academic Excellence
     u   Advancing the profession – put aside competition for the good of the profession
     u   Student Centered Education

u   Educational advancement and sustained excellence
     u   can only be achieved through collaboration,
     u   collective data sharing,
     u   and commitment to wide-spread assessment practices.

                                                                                           15
Sustained Professional Assessment

u   A crucial factor to success
     u   Development of a culture of assessment that is based
          u A sound theoretical framework

          u Integration of multiple sources of data

u   Current Professional Models

u   Barriers to national Physical Therapy educational assessment

Future Directions

                                                                   16
The Future

 u   Use existing empirical data and theoretical frameworks as a basis for
     defining excellence
 u    Reconfirm a commitment by all stakeholders to participate in
     standardized professional assessment
 u   Collaborate with CAPTE to interface data elements and collection.

CLUSTER                        ATTRIBUTE                              Descriptions
Cluster 1      Diverse and Engaged Participants                       Learning and
                               Diverse and Engaged Faculty/Students   leadership environments that
                                                                      foster student/faculty growth
                               Engaged Leaders                        and development
Cluster 2      Participatory Cultures
                               Shared Program Direction
                                                                      Program innovation through
                               Community of Learners
                                                                      collaborative interaction
                               Risk Taking/Interactive Learning
Cluster 3      Interactive Teaching and learning
                               Mentoring/Cooperative Learning         High impact practices that foster
                               Experiential Learning                  student learning
Cluster 4      Connected Program Requirements
                               Depth/Breadth Coursework               Curricular design that promotes
                               Professional Residency                 professional development
                               Tangible Products
                               Support for Faculty
Cluster 5      Adequate Resources
                               Infrastructure

                                                                                                          17
The Future

u   Require assessment participation as a professional expectation for being “a member in good
    standing” in ACAPT.
u   Designate appropriate resources in assessment activities for data collection, storage, analysis
    technology, personnel, and communication.
u   Educate constituents on VALUE and ways data can be useful for program benchmarking,
    quality improvement, strategic planning and accreditation self-study.

                                                                                                      18
The Future

u   The BenEx Taskforce envisions the development of a Research Center
     u   designed to support ACAPT initiatives,
     u   provide assistance with data collection and interpretation
     u   to guide ongoing growth and development of our educational enterprises.

u   Profession needs to support assessment
     u   Elevate programs beyond level set by CAPTE
     u   Support programmatic need to substantiate improvement

Assessment is Necessary for Excellence

u   Call to action from Jensen et al., 2018:
     u   ”Create a national data set that includes essential metrics of performance
         outcomes, structures, and processes that can be used for meaningful research
         to guide future evidence-based change”

u   An excellent academic program demonstrates a culture of excellence by
    continually and intentionally striving to transform learners, advance
    knowledge, and improve societal health.

                                                                                        19
ACAPT Criteria
for Excellence

                 20
Questions?

             References

u   Busche G. The Appreciative Inquiry Model. In: Kessler EH, ed. Encyclopedia of Management Theory. Thousand Oaks, CA: Sage Publications; 2013.

u   Grignon TP, Henley E, Lee KM, Abentroth MJ, Jette DU. Expected graduate outcomes in US physical therapist education programs: a qualitative study. J
    Phys Ther Educ. 2014;28(1):48.

u   Haworth JG, Conrad C. Emblems of quality in higher education: developing and sustaining high-quality programs. Needham, MA: Allyn and Bacon; 1997.
u   Heath, A., Altenburger, P., Brechter, J.H., Chleboun, G., Jette, D.U., Pike, G., Schilling, D., Topp, K., Tschoepe, B., 2018. J Phys Ther Educ. 32:109-117.

u   Higgs J, McMeeken J. Achieving Quality in Physiotherapy Programmes through Benchmarking. NZ J Physiother. 1997;25(3):19-22.

u   Hobson R, Rolland S, Rotgans J, et al. Quality assurance, benchmarking, assessment and mutual international recognition of qualifications. Eur J Dent Educ 2008;12(s1):92-100.
u   Jensen, G.M., Hack, L.M., Nordstrom, T., Gwyer, J., Mostrom, E. National study of excellence and innovation in physical therapist education: part 2 – a call to reform. Phys Ther. 2017.
    97:875-888.
u   Newswander LK, Borrego M. Engagement in two interdisciplinary graduate programs. Higher Education. 2009;58(4):551-562.

u   Niemczyk, M., (2012) Using your Brain to Learn: Strategies for Success, Performance Optimization Solutions LLC. Arizona State University

u   Schuwirth, L., van der Vleuten, C., Durning, S.J. What programmatic assessment in medical education can learn from healthcare. 2017. Perspect Med Educ. 6:211-215.
u   Standards and Required Elements for Accreditation of Physical Therapist Education Programs, (2016).

u   Suskie L. Assessing Student Learning: A common sense guide. San Francisco, CA: Jossey Bass; 2009.

u   Wang Hm. What matters in graduate school? Exploring patterns of student engagement, academic and personal development. Association for Institutional Research; 2003; Tampa,
    FL.

u   Warden S. Testing the Engagement Theory of Program Quality in CACREP-Accredited Counselor Education Programs. Counselor Education and Supervision. 2012;51(2):127-140.

                                                                                                                                                                                               21
You can also read