Articulate the value of this program evaluation in relation to the institutional mission.
Below are the following documents documents
3 papers you did
rubric
all to the documents you needed to do each paper
Please follow rubric verbatim
HEA 530: Final Project Guidelines and Rubric
Overview Data-driven decision-making affords higher education leadership the understanding that assessment of performance is integral in the planning, implementation, assessment, and revision of the goals that support the overall institutional mission. The ability to methodically apply empirically based, data-centric approaches to inform institutional decision-making provides the framework for monitoring content or process standards (which define what the institution must accomplish) and for monitoring performance standards (which define how well the institution must achieve what must be done). In this project, through a case study, you will compile and analyze all the relevant information to formulate hypotheses on how best to use it to achieve program goals and advance the mission of the institution.
Armed with the techniques learned in this course, you will be able to determine what data will be needed to answer institutional questions, assemble a body of empirical evidence, and arrive at decisions that support and advance the program area. In putting together your evaluation, you will begin with an executive summary that includes a timeline of the program that defines how often the evaluation would take place for audience reference; then you will present your data and the analysis of it (including any relevant graphics); you will conclude with the resulting recommendations. In addition, you will include an appendix that presents all of the data you utilized for your evaluation.
You will construct the final project incrementally throughout the course by meeting specific milestones. Support for meeting the milestones will be provided through peer feedback conducted in the Discussion Topic.
Review the requirements of the entire program evaluation below first in order to understand the scope of the project. Then, review the requirements for the milestones for completing the program evaluation and the modules in which they are due. The rubric for grading the final project is the final section of this document.
I. Executive Summary
Program Evaluation
a) Describe the organization designated in this program evaluation and the role of the functional area within the organization. Analyze how the roles and practices of various enterprise employees impact the functional area by answering these questions:
i. What is the organizational structure for the designated area? ii. What are the primarily responsibilities for each role? iii. How would you describe the academic and administrative areas that play an ancillary role in the function of this area, as well as the
assessment regimen? iv. What is the value of assessment for this functional area? v. What is the value of this program evaluation in relation to the institutional mission?
b) Articulate the value of this program evaluation in relation to the institutional mission.
II. Analyze the operational frameworks that currently drive the efforts of the program
a) Describe the current outcomes and measures of success relevant to your program. What institutional standards, milestones, benchmarks, or goals provide the framework for existing program processes and procedures? How do these frameworks help shape your program?
b) Describe any regulatory standards that govern and define successful program outcomes relevant to your program. How do these regulations help shape your program?
III. Data preparation of the information used to evaluate your program
a) Analyze your data collection process. Which existing sources inform your program? What additional or new data sources did you tap for our evaluation? Why did you select these sources to inform your evaluation? What other information would help explain your process?
b) Summarize and display your data in a standard and meaningful manner. i. Create both numerical and graphic representations of the data collected. ii. Merge the program data collected, illustrating the planned versus actual results with appropriate presentation graphics.
IV. Correlate data with desired outcomes
a) Analyze program success, considering such criteria as: i. Benchmark item: What is the specific benchmark to be met? ii. Action item: What specific action needs to be taken?
iii. Responsibility center: Who has primary responsibility for the action to be taken? iv. Others involved: What other roles or positions will be involved in the action? v. Target date: What is the targeted date of action implementation?
vi. Expected outcome: When the action is complete, what is the expected, measurable outcome? b) Analyze the impact your program in its current state is having on organizational success.
i. Benchmark item: Was the specific benchmark met? ii. Action item: If not, what specific action needs to be taken to support organizational success?
iii. Responsibility center: Who has primary responsibility for the action to be taken? iv. Others involved: What other roles or positions will be involved in the action? v. Target date: What is the targeted date of action implementation?
vi. Expected outcome: When the action is complete, what is the expected, measurable outcome relative to organizational success?
V. Make recommendations for continuous program improvement based upon the empirical data analysis a) Analyze the program for indicators that necessitate changes in program process or procedures. Was the goal met or exceeded? If not, what activities
should be performed to help identify issues/problems? What other information would be helpful in analyzing the indicators? b) Recommend actions or strategies for continuous program improvement, based upon the empirical data analysis. What continuous improvement
strategies can be implemented short-term? What continuous improvement strategies will require additional time, resources, and support? What other information would be helpful in making these recommendations?
c) Recommend improvements for future program evaluation processes and data collection efforts. What aspects of the program evaluation significantly aided in the collection and analysis of data? What aspects of the program evaluation yielded concern with regard to the collection and analysis of data, making those aspects candidates for change? What other information would be helpful in making these recommendations?
Milestones The program evaluation final project will consist of four deliverables:
Milestone One: Draft of Executive Summary In Module Three, you will share the draft of your executive summary in the Discussion Topic.
Describe the organization designated in this program evaluation and the role of the functional area within the organization. Analyze how the roles and practices of various enterprise employees impact the functional area by answering these questions:
1. What is the organizational structure for the designated area? 2. What are the primarily responsibilities for each role? 3. How would you describe the academic and administrative areas that play an ancillary role in the function of this area, as well as the assessment
regimen? 4. What is the value of assessment for this functional area?
Articulate the value of this program evaluation in relation to the institutional mission.
In Module Four, you will receive and provide constructive feedback from your classmates and your instructor through the discussion Topic.
Milestone Two: Draft of Data Presentation In Module Five, you will submit an example of a graphical representation of data that you intend to use in your program evaluation to the Discussion Topic for feedback from the instructor and classmates. Resources are provided in Module Six for students to learn how to summarize and display data.
Milestone Three: Draft of Recommendations In Module Seven, you will present your draft recommendations for continuous improvement, to include the following:
Analyze the program for indicators that necessitate changes in program process or procedures. Was the goal met or exceeded? If not, what activities should be performed to help identify issues/problems? What other information would be helpful in analyzing the indicators?
Recommend actions or strategies for continuous program improvement, based upon the empirical data analysis. What continuous improvement strategies can be implemented short-term? What continuous improvement strategies will require additional time, resources, and support? What other information would be helpful in making these recommendations?
In Module Eight, additionally, you will provide feedback to two of your classmates. Your participation in the workshop will be graded with the Peer Workshop Rubric.
Final Submission: Program Evaluation In Module Nine, you will submit your program evaluation. This submission is graded with the Final Project Rubric.
Deliverables
Milestone Deliverable Module Due Grading
1 Draft of Executive Summary Three Graded separately; Milestone One Rubric
2 Draft of Data Presentation Five Graded separately; Milestone Two Rubric
3 Draft of Recommendations Seven Graded separately; Milestone Three Rubric
Final Submission: Program Evaluation Nine Graded separately; Final Project Rubric
Final Project Rubric Guidelines for Submission: The final document will include any graphs and charts necessary and must include a timeline of the project in the executive summary. Submit your evaluation with the standard formatting: one-inch margins, Times New Roman 12-point, using the most current APA style manual for citations. In addition, you must include an appendix with all the primary data you used to compile your evaluation.
Critical Elements Exemplary (100%) Proficient (90%) Needs Improvement (70%) Not Evident (0%) Value
Executive Summary: Organization and Functional Area
Meets “Proficient” criteria, and choice of examples illustrates insight into the relationship between the organization and the area
Describes the organization and the role of the functional area within the organization
Describes the organization or the functional area, but lacks detailed discussion of the relationship between the two
Does not describe the organization or the functional area
5
Executive Summary: Employee Roles
Meets “Proficient” criteria, and details and examples illustrate insight into the relationship between the area and the employees
Analyzes the impact of the roles and practices of employees on the functional area
Discusses employee roles and practices, but impact on the functional area is overly brief and lacks detail
Does not discuss employee roles or their impact on the functional area
5
Executive Summary: Value
Meets “Proficient” criteria, and details and examples illustrate insight into the relationship between the evaluation and the mission
Articulates the value of the program evaluation in relation to the institution’s mission
Discusses the program evaluation or the institution’s mission, but lacks detail in relating the two
Does not discuss the program evaluation or the institution’s mission
5
Operational Frameworks: Current
Outcomes
Meets “Proficient” criteria, and details and examples demonstrate how the current frameworks help shape the program
Describes the current outcomes and measures of success relevant to the program
Discusses the current outcomes relevant to the program, but lacks specificity
Does not discuss current outcomes relevant to the program
5
Operational Frameworks: Regulatory Standards
Meets “Proficient” criteria, and details and examples demonstrate how the current regulatory standards help shape the program
Describes any regulatory standards that govern program outcomes relevant to the program
Discusses the relevant regulatory standards, but lacks specificity
Does not discuss relevant regulatory standards
5
Data Preparation: Collection Process
Meets “Proficient” criteria, and details and examples demonstrate a broad range of data sources and processes
Analyzes the data collection process, explaining why the sources are appropriate
Discusses the data collection process, but lacks detail in explaining why the sources are appropriate
Does not discuss the data collection process
10
Data Preparation: Summarize and
Display I
Meets “Proficient” criteria, and representations illustrate insight into the data collected
Creates detailed numerical and graphic representations of the data collected
Creates cursory numerical or graphic representations of the data collected
Does not create numerical and graphic representations of the data collected
10
Data Preparation: Summarize and
Display II
Meets “Proficient” criteria, and details and examples demonstrate facility with data presentation through the effective use of graphs, charts, and tables
Merges the program data collected, illustrating the planned vs. actual program results with appropriate presentation graphics
Merges the data illustrating the planned vs. actual program results. but use of presentation graphics is not appropriate
Does not present the data illustrating the planned vs. actual program results
10
Correlate Data: Program Success
Meets “Proficient” criteria, and data application demonstrates the relationship between the data and the program
Analyzes the program’s success by applying the data cogently and specifically
Analyzes the program’s success, but application of data is inaccurate, cursory, or overgeneralized
Does not analyze program’s success
10
Correlate Data: Organizational
Success
Meets “Proficient” criteria, and details and examples demonstrate the relationship between the program and the organization
Analyzes the impact of the program on organizational success
Discusses the impact of the program on organizational success, but is cursory or overgeneralized
Does not discuss the impact of the program on organizational success
10
Recommendations: Indicators
Meets “Proficient” criteria, and details and examples demonstrate a nuanced analysis of how the data reveals the indicators
Analyzes the program for indicators that necessitate changes in program process or procedure
Analyzes the program for indicators that necessitate change, but is inaccurate, cursory, or overgeneralized
Does not analyze the program for indicators that necessitate change
10
Recommendations: Improvement
Meets “Proficient” criteria, and details and examples ground recommendations insightfully in the data
Recommends actions or strategies for continuous program improvement based on the data
Recommends actions or strategies for continuous program improvement, but either does not base them on the data or is inaccurate, cursory, or overgeneralized
Does not recommend actions or strategies for continuous improvement
5
Recommendations: Future
Meets “Proficient” criteria, and details and examples ground recommendations insightfully in the data and the evaluation process
Recommends improvements for future program evaluation processes and data collection
Recommends improvements for future program evaluation processes and data collection, but is inaccurate, cursory, or overgeneralized
Does not recommend improvements for future program evaluation processes and data collection
5
Articulation of Response
Submission is free of errors related to citations, grammar, spelling, syntax, and organization and is presented in a professional and easy-to-read format
Submission has no major errors related to citations, grammar, spelling, syntax, or organization
Submission has major errors related to citations, grammar, spelling, syntax, or organization that negatively impact readability and articulation of main ideas
Submission has critical errors related to citations, grammar, spelling, syntax, or organization that prevent understanding of ideas
5
Earned Total 100%
Submit Your Assignment and get professional help from our qualified experts!
Articulate the value of this program evaluation in relation to the institutional mission. was first posted on July 15, 2019 at 3:13 pm.
©2019 "Submit Your Assignment". Use of this feed is for personal non-commercial use only. If you are not reading this article in your feed reader, then the site is guilty of copyright infringement. Please contact me at support@nursingassignmenttutor.com