Education

HEA 530: Final Project Guidelines and Rubric

Overview Data-driven decision-making affords higher education leadership the understanding that assessment of performance is integral in the planning, implementation, assessment, and revision of the goals that support the overall institutional mission. The ability to methodically apply empirically based, data-centric approaches to inform institutional decision-making provides the framework for monitoring content or process standards (which define what the institution must accomplish) and for monitoring performance standards (which define how well the institution must achieve what must be done). In this project, through a case study, you will compile and analyze all the relevant information to formulate hypotheses on how best to use it to achieve program goals and advance the mission of the institution.

Armed with the techniques learned in this course, you will be able to determine what data will be needed to answer institutional questions, assemble a body of empirical evidence, and arrive at decisions that support and advance the program area. In putting together your evaluation, you will begin with an executive summary that includes a timeline of the program that defines how often the evaluation would take place for audience reference; then you will present your data and the analysis of it (including any relevant graphics); you will conclude with the resulting recommendations. In addition, you will include an appendix that presents all of the data you utilized for your evaluation.

You will construct the final project incrementally throughout the course by meeting specific milestones. Support for meeting the milestones will be provided through peer feedback conducted in the Discussion Topic.

Review the requirements of the entire program evaluation below first in order to understand the scope of the project. Then, review the requirements for the milestones for completing the program evaluation and the modules in which they are due. The rubric for grading the final project is the final section of this document.

I. Executive Summary

Program Evaluation

a) Describe the organization designated in this program evaluation and the role of the functional area within the organization. Analyze how the roles and practices of various enterprise employees impact the functional area by answering these questions:

i. What is the organizational structure for the designated area? ii. What are the primarily responsibilities for each role? iii. How would you describe the academic and administrative areas that play an ancillary role in the function of this area, as well as the

assessment regimen? iv. What is the value of assessment for this functional area? v. What is the value of this program evaluation in relation to the institutional mission?

b) Articulate the value of this program evaluation in relation to the institutional mission.

II. Analyze the operational frameworks that currently drive the efforts of the program

a) Describe the current outcomes and measures of success relevant to your program. What institutional standards, milestones, benchmarks, or goals provide the framework for existing program processes and procedures? How do these frameworks help shape your program?

b) Describe any regulatory standards that govern and define successful program outcomes relevant to your program. How do these regulations help shape your program?

III. Data preparation of the information used to evaluate your program

a) Analyze your data collection process. Which existing sources inform your program? What additional or new data sources did you tap for our evaluation? Why did you select these sources to inform your evaluation? What other information would help explain your process?

b) Summarize and display your data in a standard and meaningful manner. i. Create both numerical and graphic representations of the data collected. ii. Merge the program data collected, illustrating the planned versus actual results with appropriate presentation graphics.

IV. Correlate data with desired outcomes

a) Analyze program success, considering such criteria as: i. Benchmark item: What is the specific benchmark to be met? ii. Action item: What specific action needs to be taken?

iii. Responsibility center: Who has primary responsibility for the action to be taken? iv. Others involved: What other roles or positions will be involved in the action? v. Target date: What is the targeted date of action implementation?

vi. Expected outcome: When the action is complete, what is the expected, measurable outcome? b) Analyze the impact your program in its current state is having on organizational success.

i. Benchmark item: Was the specific benchmark met? ii. Action item: If not, what specific action needs to be taken to support organizational success?

iii. Responsibility center: Who has primary responsibility for the action to be taken? iv. Others involved: What other roles or positions will be involved in the action? v. Target date: What is the targeted date of action implementation?

vi. Expected outcome: When the action is complete, what is the expected, measurable outcome relative to organizational success?

V. Make recommendations for continuous program improvement based upon the empirical data analysis a) Analyze the program for indicators that necessitate changes in program process or procedures. Was the goal met or exceeded? If not, what activities

should be performed to help identify issues/problems? What other information would be helpful in analyzing the indicators? b) Recommend actions or strategies for continuous program improvement, based upon the empirical data analysis. What continuous improvement

strategies can be implemented short-term? What continuous improvement strategies will require additional time, resources, and support? What other information would be helpful in making these recommendations?

c) Recommend improvements for future program evaluation processes and data collection efforts. What aspects of the program evaluation significantly aided in the collection and analysis of data? What aspects of the program evaluation yielded concern with regard to the collection and analysis of data, making those aspects candidates for change? What other information would be helpful in making these recommendations?

Milestones The program evaluation final project will consist of four deliverables:

Milestone One: Draft of Executive Summary In Module Three, you will share the draft of your executive summary in the Discussion Topic.

 Describe the organization designated in this program evaluation and the role of the functional area within the organization. Analyze how the roles and practices of various enterprise employees impact the functional area by answering these questions:

1. What is the organizational structure for the designated area? 2. What are the primarily responsibilities for each role? 3. How would you describe the academic and administrative areas that play an ancillary role in the function of this area, as well as the assessment

regimen? 4. What is the value of assessment for this functional area?

 Articulate the value of this program evaluation in relation to the institutional mission.

In Module Four, you will receive and provide constructive feedback from your classmates and your instructor through the discussion Topic.

Milestone Two: Draft of Data Presentation In Module Five, you will submit an example of a graphical representation of data that you intend to use in your program evaluation to the Discussion Topic for feedback from the instructor and classmates. Resources are provided in Module Six for students to learn how to summarize and display data.

Milestone Three: Draft of Recommendations In Module Seven, you will present your draft recommendations for continuous improvement, to include the following:

 Analyze the program for indicators that necessitate changes in program process or procedures. Was the goal met or exceeded? If not, what activities should be performed to help identify issues/problems? What other information would be helpful in analyzing the indicators?

 Recommend actions or strategies for continuous program improvement, based upon the empirical data analysis. What continuous improvement strategies can be implemented short-term? What continuous improvement strategies will require additional time, resources, and support? What other information would be helpful in making these recommendations?

In Module Eight, additionally, you will provide feedback to two of your classmates. Your participation in the workshop will be graded with the Peer Workshop Rubric.

Final Submission: Program Evaluation In Module Nine, you will submit your program evaluation. This submission is graded with the Final Project Rubric.

Deliverables

Milestone Deliverable Module Due Grading

1 Draft of Executive Summary Three Graded separately; Milestone One Rubric

2 Draft of Data Presentation Five Graded separately; Milestone Two Rubric

3 Draft of Recommendations Seven Graded separately; Milestone Three Rubric

Final Submission: Program Evaluation Nine Graded separately; Final Project Rubric

Final Project Rubric Guidelines for Submission: The final document will include any graphs and charts necessary and must include a timeline of the project in the executive summary. Submit your evaluation with the standard formatting: one-inch margins, Times New Roman 12-point, using the most current APA style manual for citations. In addition, you must include an appendix with all the primary data you used to compile your evaluation.

Critical Elements Exemplary (100%) Proficient (90%) Needs Improvement (70%) Not Evident (0%) Value

Executive Summary: Organization and Functional Area

Meets “Proficient” criteria, and choice of examples illustrates insight into the relationship between the organization and the area

Describes the organization and the role of the functional area within the organization

Describes the organization or the functional area, but lacks detailed discussion of the relationship between the two

Does not describe the organization or the functional area

5

Executive Summary: Employee Roles

Meets “Proficient” criteria, and details and examples illustrate insight into the relationship between the area and the employees

Analyzes the impact of the roles and practices of employees on the functional area

Discusses employee roles and practices, but impact on the functional area is overly brief and lacks detail

Does not discuss employee roles or their impact on the functional area

5

Executive Summary: Value

Meets “Proficient” criteria, and details and examples illustrate insight into the relationship between the evaluation and the mission

Articulates the value of the program evaluation in relation to the institution’s mission

Discusses the program evaluation or the institution’s mission, but lacks detail in relating the two

Does not discuss the program evaluation or the institution’s mission

5

Operational Frameworks: Current

Outcomes

Meets “Proficient” criteria, and details and examples demonstrate how the current frameworks help shape the program

Describes the current outcomes and measures of success relevant to the program

Discusses the current outcomes relevant to the program, but lacks specificity

Does not discuss current outcomes relevant to the program

5

Operational Frameworks: Regulatory Standards

Meets “Proficient” criteria, and details and examples demonstrate how the current regulatory standards help shape the program

Describes any regulatory standards that govern program outcomes relevant to the program

Discusses the relevant regulatory standards, but lacks specificity

Does not discuss relevant regulatory standards

5

Data Preparation: Collection Process

Meets “Proficient” criteria, and details and examples demonstrate a broad range of data sources and processes

Analyzes the data collection process, explaining why the sources are appropriate

Discusses the data collection process, but lacks detail in explaining why the sources are appropriate

Does not discuss the data collection process

10

Data Preparation: Summarize and

Display I

Meets “Proficient” criteria, and representations illustrate insight into the data collected

Creates detailed numerical and graphic representations of the data collected

Creates cursory numerical or graphic representations of the data collected

Does not create numerical and graphic representations of the data collected

10

Data Preparation: Summarize and

Display II

Meets “Proficient” criteria, and details and examples demonstrate facility with data presentation through the effective use of graphs, charts, and tables

Merges the program data collected, illustrating the planned vs. actual program results with appropriate presentation graphics

Merges the data illustrating the planned vs. actual program results. but use of presentation graphics is not appropriate

Does not present the data illustrating the planned vs. actual program results

10

Correlate Data: Program Success

Meets “Proficient” criteria, and data application demonstrates the relationship between the data and the program

Analyzes the program’s success by applying the data cogently and specifically

Analyzes the program’s success, but application of data is inaccurate, cursory, or overgeneralized

Does not analyze program’s success

10

Correlate Data: Organizational

Success

Meets “Proficient” criteria, and details and examples demonstrate the relationship between the program and the organization

Analyzes the impact of the program on organizational success

Discusses the impact of the program on organizational success, but is cursory or overgeneralized

Does not discuss the impact of the program on organizational success

10

Recommendations: Indicators

Meets “Proficient” criteria, and details and examples demonstrate a nuanced analysis of how the data reveals the indicators

Analyzes the program for indicators that necessitate changes in program process or procedure

Analyzes the program for indicators that necessitate change, but is inaccurate, cursory, or overgeneralized

Does not analyze the program for indicators that necessitate change

10

Recommendations: Improvement

Meets “Proficient” criteria, and details and examples ground recommendations insightfully in the data

Recommends actions or strategies for continuous program improvement based on the data

Recommends actions or strategies for continuous program improvement, but either does not base them on the data or is inaccurate, cursory, or overgeneralized

Does not recommend actions or strategies for continuous improvement

5

Recommendations: Future

Meets “Proficient” criteria, and details and examples ground recommendations insightfully in the data and the evaluation process

Recommends improvements for future program evaluation processes and data collection

Recommends improvements for future program evaluation processes and data collection, but is inaccurate, cursory, or overgeneralized

Does not recommend improvements for future program evaluation processes and data collection

5

Articulation of Response

Submission is free of errors related to citations, grammar, spelling, syntax, and organization and is presented in a professional and easy-to-read format

Submission has no major errors related to citations, grammar, spelling, syntax, or organization

Submission has major errors related to citations, grammar, spelling, syntax, or organization that negatively impact readability and articulation of main ideas

Submission has critical errors related to citations, grammar, spelling, syntax, or organization that prevent understanding of ideas

5

Earned Total 100%

Order now and get 10% discount on all orders above $50 now!!The professional are ready and willing handle your assignment.

ORDER NOW »»