Connecting the Pieces: A Guide to Creating an Effective PEP

Similar documents
Connecting the Pieces: A Guide to Creating an Effective PEP

PROGRAM EFFECTIVENESS PLAN GUIDEBOOK

Last Review: Outcome: Next Review:

Guidelines: Preparation of the Focused Site Visit Report 2017 Standards and Criteria

Standards for Accreditation of. Baccalaureate and. Nursing Programs

1. Text in red are additions. 2. Text high-lighted in yellow with strikeout are deletions.

ALABAMA BOARD OF NURSING ADMINISTRATIVE CODE CHAPTER 610-X-3 NURSING EDUCATION PROGRAMS TABLE OF CONTENTS

1.1 The mission/philosophy and outcomes of the nursing education unit are congruent with those of the governing organization.

Time/ Frequency of Assessment. Person Responsible. Associate Dean and Program Chair. Every 3 years Or accompanying curriculum change

FUNDING APPLICATION RFP For Former OJJDP Funded YouthBuild Affiliated Programs OJJDP Mentoring Funding Due: October 31, 2014

GUIDELINES FOR THE PREPARATION OF THE SELF-STUDY REPORT UTILIZING THE 2013 ACEN STANDARDS AND CRITERIA

CHAPTER VII ST PROGRAMMATIC EVALUATION STANDARDS FOR SURGICAL TECHNOLOGY

DO NOT MAIL THIS APPLICATION OR ANY MATERIAL TO THE SEANC CENTRAL OFFICE.

Ark. Admin. Code I Alternatively cited as AR ADC I. Vision Statement

ACEN 2013 STANDARDS AND CRITERIA MASTER S and POST-MASTER S CERTIFICATE

September 2013 COMPARISON CROSSWALK PRE-LICENSURE NURSING PROGRAMS CABRN-CCNE

1. Text in red are additions. 2. Text high-lighted in yellow with strikeout are deletions.

CCNE Standard I: Program Quality: Mission and Governance

CCNE Standard I: Program Quality: Mission and Governance

LAC 46: XLVII STANDARDS AND REQUIREMENTS FOR GRADUATE NURSING EDUCATION DEGREE PROGRAMS: MISSION/PHILOSOPHY AND GOALS

CANDIDACY: GUIDE FOR PRESENTATION Candidacy Status

SITE VISIT REPORT «Governing_Organization» «CEO_City», «CEO_State» Instructions: Verify accuracy for all pre-populated General Information.

South Carolina Nursing Education Programs August, 2015 July 2016

SYSTEMATIC PLAN of EVALUATION FOR RSU RN-to-BSN PR0GRAM FY Standard 6

STATUTORY AUTHORITY: Nursing Practice Act, Section NMSA 1978 Comp. [ NMAC - Rp,

SITE VISIT REPORT «Governing_Organization» «CEO_City», «CEO_State» Instructions: Verify accuracy for all pre-populated General Information.

Program Outcomes Summary BSN Program % Kaplan IT system

DEPARTMENT OF LICENSING AND REGULATORY AFFAIRS DIRECTOR S OFFICE NURSING HOME ADMINISTRATORS GENERAL RULES

COVER SHEET. 1. Institution Name. 2. State. 3. Date submitted MM DD YYYY / / 4. Report Preparer's Information: Name of Preparer: ( ) -

MASTER PLAN OUTCOMES EVALUATION BSN PROGRAM

Annual Report. Instructions

Ohio Means Internships & Co-ops 4 Request for Proposals Application Release: 2/22/17 Application Due: 3/22/17

LAC 46:XLVII MISSION/PHILOSOPHY AND GOALS

TROY School of Nursing Evaluation Plan. Assessment Method/s

ALABAMA BOARD OF NURSING ALL OUT OF STATE INSTITUTIONS OFFERING NURSING PROGRAMS IN ALABAMA

Webinar Control Panel

Statewide to: Technical Center Directors

Section C: Standards for Programmatic Accreditation

DEPARTMENT OF LICENSING AND REGULATORY AFFAIRS DIRECTOR S OFFICE MASSAGE THERAPY - GENERAL RULES PART 1. GENERAL RULES

CSUSB STEM Teaching Scholarship

34 CFR 690 Federal Pell Grant Program

6.1 ELA: The Systematic Plan for Evaluation will include all of the following data with discussion of results and action for development

34 CFR 690. Integrated Regulations Incorporating. Program Integrity Issues Final Rules (published in October 29, 2010 Federal Register)

MASTER S/PMC. STANDARD 1 Mission and Administrative Capacity COMMENTS Mission and Administrative Capacity

Western Kentucky University School of Nursing. Faculty/Staff Handbook

Outputs Outcomes -- Impact Activities Participation Process (what & when) Impact Outcome

Experiential Education

DEPARTMENT OF LICENSING AND REGULATORY AFFAIRS DIRECTOR S OFFICE BOARD OF NURSING - GENERAL RULES

Quality Improvement Work Plan

COMPREHENSIVE COUNSELING INITIATIVE FOR INDIANA K-12 STUDENTS REQUEST FOR PROPOSALS COUNSELING INITIATIVE ROUND II OCTOBER 2017

State Board of Technical Professions

BACCALAUREATE. STANDARD 1 Mission and Administrative Capacity COMMENTS Mission and Administrative Capacity

Assessment of the Associate Degree Nursing Program St. Charles Community College Academic Year

2015 Associations Matter Study Interim Results

MASTER OF SCIENCE IN NURSING: COMMUNITY AND PUBLIC HEALTH NURSING SPECIALIZATION

ASSOCIATE. STANDARD 1 Mission and Administrative Capacity COMMENTS Mission and Administrative Capacity. Mission and Administrative Capacity

2017 Educator Preparation Performance Report Multi Age (P-12) Teaching English to Speakers of Other Languages (TESOL) Ohio State University

Medicaid Managed Specialty Supports and Services Concurrent 1915(b)/(c) Waiver Program FY 17 Attachment P7.9.1

The University of North Carolina at Chapel Hill School of Nursing. Comprehensive Examination Requirements

Two-year Associate Degree Nursing (ADN) Program RN Applicant Checklist

Cal Grant Workshop Beginner Track

Institutional Assessment Report

Using Your PSAT/NMSQT Scores to Build College Readiness

Guidelines on continuing professional development

Request for Proposals. For. Clinical Skills Assessment Pilot Project

Military recruiting expectations for homeschooled graduates compiled, April 2010

ASSOCIATE PREVENTION SPECIALISTS (APS)

REGISTERED NURSING PROGRAM RN INFORMATION PACKET

Developmental Disabilities Nurses Association

Eastern Michigan University Clinical Mental Health Counseling College Counseling School Counseling Program Evaluation April 2017

Accreditation Council for Occupational Therapy Education (ACOTE ) Standards and Interpretive Guidelines August 2012

ACEN 2013 STANDARDS AND CRITERIA CLINICAL DOCTORATE/DNP SPECIALIST CERTIFICATE

VET Student Handbook

Ohio Common Grant Form GRANT APPLICATION SHORT FORM

Contents. Page 1 of 42

Diagnostic Waiting Times

Radiologic Technology Program (AAS RDTN) Assessment Plan Narrative. Program Outcomes:

Accreditation Commission for Midwifery Education

N.J.A.C. 6A:19, Career and Technical Education Programs and Standards. Summary

ROLES & RESPONSIBILITIES

COUNSELING CREDENTIALS

COMPARISON CROSSWALK '

Michigan Department of Community Health Diabetes Self-Management Education Program Standards

Strategic Employment Placement Grant

ASSOCIATE DEGREE NURSING PROGRAM DIRECTOR Job Description. Commensurate with Education and Experience

Community Health Centre Program

ASSOCIATE PREVENTION SPECIALISTS (APS)

Preceptor Guidelines and Application

The University of North Carolina at Chapel Hill School of Nursing. Comprehensive Examination Requirements

Scholarly Project Handbook Doctor of Nursing Practice Program

OUTCOME TRENDS IN NURSING EDUCATION Updated February 15, 2018

This policy shall apply to all directly-operated and contract network providers of the MCCMH Board.

APPLICATION FOR A BACCALAUREATE DEGREE PROGRAM

TITLE I 201 PLANNING, INITIATING AND COMPLETING YOUR COMPREHENSIVE NEEDS ASSESSMENT (CNA)

kaiser medicaid and the uninsured commission on O L I C Y

DEPARTMENT OF LICENSING AND REGULATORY AFFAIRS DIRECTOR S OFFICE BOARD OF NURSING - GENERAL RULES. Filed with the Secretary of State on

New Jersey State Legislature Office of Legislative Services Office of the State Auditor. July 1, 2011 to September 7, 2016

ROAD TO INDEPENDENCE PROGRAM REINSTATEMENT APPLICATION

NURSING INFORMATION AND ENROLLMENT PACKET FOR SUMMER 2018 LVN-TO-RN CAREER MOBILITY PROGRAM

CHAPTER SIX STANDARDS FOR NURSING EDUCATION PROGRAMS

Transcription:

Program Effectiveness Plan Outline Connecting the Pieces: A Guide to Creating an Effective PEP This outline has been created to assist schools in developing an PEP reflecting the areas outlined in the Accreditation Manual. Retention Mgmt Credentialing Objectives Faculty Students Grads Curriculum Mission Clinical Externships Placement Employers Surveys? 7777 Leesburg Pike Suite 314 North Falls Church, Virginia Phone: 703-917-9503 Fax: 703-917-4109 Website: www..org

The Program Effectiveness Plan The Program Effectiveness Plan (PEP) is an internal quality assessment tool that is used for evaluating each program by designing strategies to improve performance within an institution by: Identifying historical and current outcomes; Establishing and documenting specific goals; and Creating strategies to meet such goals The process of developing and maintaining a PEP requires that an institution use its past and present performance to set goals for its future performance. The PEP continuously assesses: Where have we been? Baseline data provides a foundation for gauging and demonstrating improvements. Where are we now? Current data is compared to past data to determine the areas needing improvement Where do we want to go? Future goals are established to bring about improvement in the program s effectiveness Develop the processes and strategies that will be used to achieve these new goals. Program Effectiveness Plan Outline Page 1

Developing a PEP involves collecting, maintaining, and using information about each program that reflects the specific areas outlined in Chapter V, Section I of the Accreditation Manual. The data should be analyzed for the 12-month reporting period, July 1 through June 30, and used as the foundation for making comparisons across future reporting periods. The PEP must be updated at least annually. The PEP is unique to each program and institution. An institution must demonstrate its efforts to ensure continuous improvement on an annual basis by: Systematically collecting data and information on each of the student achievement indicators; Completing an analysis of the data and information, including but not limited to, performing a comparison with previous years data; and Identifying, based upon the analysis of the data, what changes in strategies and/or activities will be made to increase program effectiveness. - July 1 June 30 AR Report Dec Jan Feb Advisory Board Meeting Oct Nov March April PEP Review Sept May Compiling & Summarizing Data Aug June End of Year Program Effectiveness Plan Outline Page 2

PEP Cover Page reporting period covered by the following PEP (July 1, 20XX through June 30, 20XX): Prepared By: Date Reviewed: For the purposes of this outline, the following is a suggested title page for an individual PEP: Name of Institution ID Code (Renewal Applicants Only) Street Address: City: State: Zip: Telephone # Website The information provided in the chart below must match that approved by the appropriate oversight agency(ies) and is consistent with other institutional publications. Program Title (As approved by oversight agency(ies) listed in question #7) PROGRAM INFORMATION Number of Instructional Weeks per day (D), evening (E), & weekend (W), if applicable Example: 40-D, 50-E, 60-W In Class Clock Hours *Outside Class Hours Total Clock Hours Identify the # of credits offered for each applicable program CHECK ONE Quarter Semester Credential Awarded by institution upon program completion Example: Diploma, Certificate, or Type of Degree (Do not use abbreviations) NOTE: This field is not applicable to all programs. Please review Standard, IV.G.2 of the Accreditation Manual for details regarding Outside Class Hours that affect and/or are applied to the total program length. Institutions awarding credit for outside class hours will be required to provide a detailed analysis of how these hours were derived, how they complement the given coursework, and how student s benefit from the respective assignments. If any portion of a program(s) is offered via distance education, complete the table below: *Blended and/or Full Program Title (As approved by oversight agency(ies) listed in question #7) Distance Education Delivery Types of Courses [e.g., general education, core, remote lab/externship/clinical] Program Effectiveness Plan Outline Page 3

Program Effectiveness Plan Content V.I.2. A program has an established documented plan for assessing its effectiveness as defined by specific outcomes. While each program must represent each element required below, the plan may be a comprehensive one which collectively represents all programs within the institution or may be individual plans for each distinct program. The Program Effectiveness Plan clearly describes the following elements: a. Student population b. Program objectives c. Retention rate d. Job placement rate e. Credentialing examination participation rate f. Credentialing examination pass rate g. Surveys that measure (i) participation, and (ii) satisfaction for the following: 1. Students (classroom and clinical experience); 2. Clinical extern affiliates; 3. Graduates; and 4. Employers PEP Introduction The first two elements, student population and program objectives, provide background information to support the PEP. Program Effectiveness Plan Outline Page 4

V.I.2.a: Student Population A description of the characteristics of the student population is included in the Plan. The institution must determine which student characteristics will be most valuable to use for assessing program effectiveness as it relates to the required elements of the PEP. This information could be identified in a list, a chart, or narrative format. The following is a list of student population demographics an institution might consider when describing its student population: Gender Estimated Family Contribution (EFC) Race Married/Single Age High School Graduate/GED Full-time/Part-time Earned average or above GPA in HS Employed/Unemployed English/Math First time college/had prior English as a second language postsecondary education Income level NOTE: The institution should choose carefully and refer back to its demographics when analyzing specific elements of the PEP. V.I.2.b: Program Objectives Program objectives are to be included as part of the PEP and must be consistent with all other institutional documents describing the program. An annual review is conducted to assess whether the program objectives are consistent with the field of study and the credential(s) awarded to include the comprehensive preparation of graduates to work in the career field. Program Effectiveness Plan Outline Page 5

V.I.2.c: Retention Rate A program demonstrates that students complete their program (V.I.1.a.). The retention rate is determined by using the required method of calculation, for the reporting period July 1 through June 30, as follows: (EE + G) / (BE + NS + RE) = R% EE= Ending Enrollment (as of June 30) G= Graduates BE= Beginning Enrollment (as of July 1) NS= New Starts RE= Re-Entries R% = Retention Percentage The retention rate for the previous two years and the current year is identified. This is determined by using the required method of calculation for the reporting period July 1 through June 30. Based upon these rates, the institution must then identify its retention rate goal for the next reporting year, the factors considered in determining such a goal, and the activities undertaken to meet the goal. Program Name & Credential (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. BACK-UP DOCUMENTATION At a minimum, an institution maintains the names of all enrollees by program, start date, and graduation date. This supporting documentation is not required to be included in the PEP, but must be readily available upon request. Program Effectiveness Plan Outline Page 6

Outcomes assessment Retention V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding RETENTION. 1. What is the process for tracking retention data? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? 2. How frequently is retention data reviewed? 3. Are there any trends apparent in the retention data? If so, what are the contributing factors? 4. What changes have been made based upon the analysis of the retention data? 5. What factors were taken into consideration in determining the retention goal? 6. What activities will be undertaken to meet the retention goal? Program Effectiveness Plan Outline Page 7

V.I.2.d: Job Placement Rate A program demonstrates that graduates are successfully employed in the field, or related field, for which they were trained (V.I.1.d.). An institution has a system in place to assist with the successful initial employment of its graduates. A graduate must be employed for 15 days and the verification must take place no earlier than 15 days after employment. The placement rate is determined by using the required method of calculation, for the reporting period July 1 through June 30, as follows: Placement Rate = (F + R)/(G-U) F = Graduates placed in their field of training R* = Graduates placed in a related field of training G = Total graduates U** = Graduates unavailable for placement *Related field refers to a position wherein the graduate s job functions are related to the skills and knowledge acquired through successful completion of the training program. **Unavailable is defined only as documented: health-related issues, military obligations, incarceration, continuing education status, or death. Important Note: graduates pending required credentialing/licensure in a regulated profession required to work in the field and, thus, not employed or not working in a related field as defined above, should be reported through back-up information required in the Annual Report. This fact will then be taken into consideration if the program placement rate falls below expectations and an Action Plan is required by. BACK-UP DOCUMENTATION At a minimum, an institution maintains the names of graduates, place of employment, job title, employer telephone numbers, and employment and verification dates. For any graduates identified as self-employed, an institution maintains evidence of employment. For any graduates identified as unavailable, the reason must be stated. Documentation in the form of employer or graduate verification forms or other evidence of employment is retained. An institution has a system in place to assist with the successful initial employment of its graduates. The institution is required to verify the employment post-initial employment date no earlier than 15 days after employment. Program Effectiveness Plan Outline Page 8

The job placement rate for the previous two years and the current year is identified. This is determined by using the required method of calculation for the reporting period July 1 through June 30. Based upon these rates, the institution must then identify its placement rate goal for the next reporting year, the factors considered in determining such a goal, and the activities undertaken to meet the goal. Program Name & Credential (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. Program Effectiveness Plan Outline Page 9

Outcomes assessment Job Placement V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding JOB PLACEMENT. 1. What is the process for tracking job placement data? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? 2. How frequently is job placement data reviewed? 3. Are there any trends apparent in the job placement data? If so, what are the contributing factors? 4. What changes have been made based upon the analysis of the job placement data? 5. What factors were taken into consideration in determining the job placement goal? 6. What activities will be undertaken to meet the job placement goal? Program Effectiveness Plan Outline Page 10

V.I.2.e: Credentialing Examination Participation Rate & V.I.2.f: Credentialing Examination Pass Rate These sections are being combined given the relationship of the credentialing examination participation and pass rates. A program demonstrates that graduates participate on credentialing exams required for employment (V.I.1.b.). If a license or credential is required (i) for employment within the geographic area served by the institution (ii) by regulatory bodies (e.g., state or other governmental agencies), or (iii) by the programmatic accrediting body, then the participation of program graduates in credentialing or licensure examinations is monitored and evaluated. The credentialing participation rate is determined by using the required method of calculation, for the reporting period July 1 through June 30, as follows: Examination participation rate = GT/GE GT = Total graduates taking examination GE= Total graduates eligible to sit for examination A program demonstrates that graduates are successful on credentialing examinations required for employment (V.I.1.c.). If an institution or program is required to monitor participation rates, then it must review graduate success on credentialing and/or licensing examinations. This review includes curricular areas in need of improvement. A program maintains documentation of such review and any pertinent curricular changes made as a result. The credentialing pass rate is determined by using the required method of calculation, for the reporting period July 1 through June 30, as follows: Examination Pass Rate = GP/GT GP = Graduates passing examination (any attempt) GT = Total graduates taking examination Program Effectiveness Plan Outline Page 11

The following screening questions are based upon the criteria per Standard, V.I.1.b, and are provided to assist the institution is determining whether it needs to complete and include these sections within the PEP: Screening Question #1: Is a license or credential required for graduates of the program to obtain employment within the geographic area served by the institution? If YES, then the following sections must be included within your PEP. If No, proceed to Screening Question #2 Screening Question #2: Is this program governed by a regulatory body (e.g., state or other governmental agencies) that requires program graduates to participate in licensure or credentialing examinations? If YES, then the following sections must be included within your PEP. If No, proceed to Screening Question #3. Screening Question #3: Is this program accredited by another agency that requires program graduates to participate in licensure or credentialing examinations? If YES, then the following sections must be included within your PEP. If No, then state in your PEP that these sections do not apply to your program and proceed to the next section. BACK-UP DOCUMENTATION At a minimum, the names of all graduates by program, actual graduation date, and the credentialing or licensure exam for which they are required to sit for employment are maintained. Program Effectiveness Plan Outline Page 12

The credentialing examination participation rate for the previous two years and the current year is identified. This is determined by using the required method of calculation for the reporting period July 1 through June 30. Based upon these rates, the institution must then identify its credentialing participation rate goal for the next reporting year, the factors considered in determining such a goal, and the activities undertaken to meet the goal. Participation Rate(s): Program Name & Credential Ex. Medical Assistant Diploma Pass Rate(s): (2 years prior) (1 year prior) (current) (goal) Program Name & Credential (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. Program Effectiveness Plan Outline Page 13

Outcomes assessment Credentialing Examination Participation & Pass Rates V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding CREDENTIALING. 1. What is the process for tracking credentialing data? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? A. Examination Participation Rates B. Examination Pass Rates 2. How frequently is credentialing data reviewed? A. Examination Participation Rates B. Examination Pass Rates 3. Are there any trends apparent in the credentialing data? If so, what are the contributing factors? A. Examination Participation Rates B. Examination Pass Rates 4. What changes have been made based upon the analysis of the credentialing data? A. Examination Participation Rates B. Examination Pass Rates 5. What factors were taken into consideration in determining the credentialing goal? A. Examination Participation Rates B. Examination Pass Rates 6. What activities will be undertaken to meet the credentialing goal? A. Examination Participation Rates B. Examination Pass Rates Program Effectiveness Plan Outline Page 14

V.I.2.g: Surveys for Students (classroom and clinical experience); Clinical Extern Affiliates; Graduates & Employers i. A program demonstrates that its required constituencies participate in completing program surveys (V.I.1.e.). A program must survey the following: 1) current students (classroom and clinical experience); 2) clinical extern affiliates; 3) graduates; and 4) employers. The purpose of the surveys is to collect data regarding a perception of a program s strengths and weaknesses. The survey participation rate is determined by using the required method of calculation, for the reporting period July 1 through June 30, as follows: Survey Participation Rate = SP / NS SP = Survey Participation (those who actually filled out the survey) NS = Number Surveyed (total number of surveys sent out) ii. A program demonstrates that it has developed survey satisfaction benchmarks based on required constituency surveys (V.I.1.f.). A program must establish satisfaction benchmarks for current students (classroom and clinical experiences), clinical extern affiliates, graduates, and employers. The purpose of the benchmarks is to collect data regarding satisfaction with the program s stated objectives and goals. The benchmark satisfaction rate is determined by using the required method of calculation, for the reporting period July 1 through June 30, as follows: Benchmark Satisfaction Rate = SL/SP SL = Satisfaction level SP = Survey Participation At a minimum, an annual review of the results is conducted and shared with administration, faculty and advisory boards. Decisions and action plans are based upon the review of the surveys, and any changes made are documented (e.g., meeting minutes, memoranda). Program Effectiveness Plan Outline Page 15

Accordingly, a program must document that at a minimum the survey data included in its effectiveness assessment include the following: Student (classroom and clinical experiences): Student surveys provide insight regarding student satisfaction relative to all aspects of the program such as instruction, educational resources, and student services, as well as their clinical experience. The surveys identify strengths and weaknesses from a student s perspective. NOTE: An institution must evidence that it obtains student feedback regarding both classroom and clinical experiences. This can be accomplished by an institution utilizing one comprehensive survey given at the end of the program, or utilizing multiple surveys administered at different times throughout the program. Clinical Extern Affiliate: Externship site surveys include the following: (i) a critique of students knowledge and skills upon completion of their in-school training; (ii) an evaluation of how well the students are trained to perform their required tasks; (iii) an assessment of the strengths and weaknesses, and proposed changes, in the instructional activities for currently enrolled students; and (iv) an assessment of the responsiveness and support provided by the designated school representative, who visited the site and remained in contact with the site throughout the duration of the students externship. Graduate: A program has a systematic plan for regularly surveying graduates, which determines if: (i) graduates have been informed of applicable credentialing requirements; (ii) the classroom, laboratory, and clinical experiences prepared students for employment; and (iii) graduates are satisfied with their educational training. Employer: A program has a systematic plan for regularly surveying employers, which determines if: (i) information on whether the skill level of the employee is adequate, and (ii) if the employer would hire another graduate from the program. Student, clinical extern affiliate, graduate, and employer surveys must include questions/content that will evidence the items detailed in the satisfaction benchmarks identified above. Program Effectiveness Plan Outline Page 16

STUDENT SURVEYS: The institution establishes: 1) a goal for the percent of surveys returned and 2) benchmarks for the level of satisfaction desired. Student (Classroom) Participation Rate(s): Program Name & Credential Ex. Medical Assistant Diploma (2 years prior) Student (Classroom) Satisfaction Rate(s): Program Name & Credential (1 year prior) (current) (goal) Ex. Medical Assistant Diploma (2 years prior) Student (Clinical Experiences) Participation Rate(s): Program Name & Credential (1 year prior) (current) (goal) Ex. Medical Assistant Diploma (2 years prior) Student (Clinical Experiences) Satisfaction Rate(s): Program Name & Credential (1 year prior) (current) (goal) (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial applicants and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. Program Effectiveness Plan Outline Page 17

Outcomes assessment Student Surveys V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding STUDENT SURVEYS. 1. What is the process for tracking student survey data? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? 2. How frequently is student survey data reviewed? 3. Are there any trends apparent in the student survey data? If so, what are the contributing factors? 4. What changes have been made based upon the analysis of the student survey data? 5. What factors did you take into consideration in determining the student survey goal or benchmark? 6. What activities will be undertaken to meet the student survey goal or benchmark? Program Effectiveness Plan Outline Page 18

CLINICAL EXTERN AFFILIATE SURVEYS: The institution establishes: 1) a goal for the percent of surveys returned and 2) benchmarks for the level of satisfaction desired. Clinical Affiliate Participation Rate(s): Program Name & Credential Ex. Medical Assistant Diploma (2 years prior) Clinical Affiliate Satisfaction Rate(s): Program Name & Credential (1 year prior) (current) (goal) (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. Program Effectiveness Plan Outline Page 19

Outcomes assessment Clinical Extern Affiliate Surveys V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding CLINICAL EXTERN AFFLIATE SURVEYS. 1. What is the process for tracking clinical extern affiliate survey data? Why is data tracked in that manner? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? 2. How frequently is clinical extern affiliate survey data reviewed? 3. Are there any trends apparent in the clinical extern affiliate survey data? If so, what are the contributing factors? 4. How is this clinical extern affiliate survey data used to improve the educational process? 5. What factors did you take into consideration in determining the clinical extern affiliate survey goal or benchmark? 6. What activities will be undertaken to meet the clinical extern affiliate survey goal or benchmark? Program Effectiveness Plan Outline Page 20

GRADUATE SURVEYS: The institution establishes: 1) a goal for the percent of surveys returned and 2) benchmarks for the level of satisfaction desired. Graduate Survey Participation Rate(s): Program Name & Credential Ex. Medical Assistant Diploma (2 years prior) Graduate Survey Satisfaction Rate(s): Program Name & Credential (1 year prior) (current) (goal) (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. Program Effectiveness Plan Outline Page 21

Outcomes assessment Graduate Surveys V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding GRADUATE SURVEYS. 1. What is the process for tracking graduate survey data? Why is data tracked in that manner? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? 2. How frequently is the graduate survey data reviewed? 3. Are any trends apparent in the graduate survey data? If so, what are the contributing factors? 4. What changes have been made based upon the graduate survey data? 5. What factors were taken into consideration in determining the graduate survey goal or benchmark? 6. What activities will be undertaken to meet the graduate survey goal or benchmark? Program Effectiveness Plan Outline Page 22

EMPLOYER SURVEYS: The institution establishes: 1) a goal for the percent of surveys returned and 2) benchmarks for the level of satisfaction desired. Employer Survey Participation Rate(s): Program Name & Credential Ex. Medical Assistant Diploma (2 years prior) Employer Survey Pass Rate(s): Program Name & Credential (1 year prior) (current) (goal) (2 years prior) (1 year prior) (current) (goal) Ex. Medical Assistant Diploma *Initial and/or new program applicants are only required to report data that is available at the time of the PEP submission as part of the Self Evaluation Report (SER); however, updated information must be available at the time of an on-site evaluation visit. Program Effectiveness Plan Outline Page 23

Outcomes assessment Employer Surveys V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of each program s effectiveness. Responding to each of the following questions will assist the institution in conducting an assessment of each program s effectiveness regarding EMPLOYER SURVEYS. 1. What is the process for tracking employer survey data? Why is data tracked in that manner? Who (this might be one or more individuals) is responsible for collecting and monitoring the data? 2. How frequently is employer survey data reviewed? 3. Are there any trends apparent in the employer survey data? If so, what are the contributing factors? 4. What changes have been made based upon the employer survey data? 5. What factors were taken into consideration in determining the employer survey goal or benchmark? 6. What activities will be undertaken to meet the employer survey goal or benchmark? Program Effectiveness Plan Outline Page 24

Curriculum Assessment V.I.3. A program has a process for assessing effectiveness annually. The Program Effectiveness Plan specifies a process and a timetable for the annual assessment of program effectiveness. The plan must include an assessment of the curriculum that uses tools which might include examinations, advisory board input, competency and skill outcomes, faculty review of resource materials, and graduate and employer surveys. Results of the assessment are not required to be reported to, but are considered in annual curriculum revision by such parties as the program supervisor, faculty, and the advisory board. Changes adopted are included in the program effectiveness plan. Responding to each of the following questions will assist the institution completing a CURRICULUM ASSESSMENT. 1. What is the process for assessing program curriculum? Who (this might be one or more individuals) is responsible for assessing program curriculum? 2. How frequently is the curriculum assessed? 3. What strengths and/or weaknesses were identified? 4. What changes have been made to improve the curriculum and/or educational process? Program Effectiveness Plan Outline Page 25

Summary & Reflection In reviewing the outcomes data and analyzing the results, the PEP should identify any changes in educational operations and/or activities the program will or has made based upon its analysis. A program might consider describing its plans for the future growth and development of such things as facilities and equipment. The results of a PEP are never final! Program Effectiveness Plan Outline Page 26