- To learn about the importance of working together with staff and families to evaluate the program.
- To learn about the importance of using program evaluation data to improve program outcomes for children and families.
- To learn about the importance of modeling continuous quality improvement so that staff focus on program goals and outcomes.
Learn
Teach
In this lesson, you will learn about the importance of program evaluation as a source for determining what services to children and families are working well and what aspects of the program are in need of improvement. All staff and families may participate in the program evaluation. Program evaluation is typically completed on an annual basis in order to determine whether the program is effectively meeting its goals. The findings of the program evaluation are typically shared with all of the program’s stakeholders: families, staff, and in some cases the community. In military child and youth programs these reports are also shared with the military leadership or higher headquarters within your Service.
The management team works with staff and families to collect relevant data, analyze it, and use the findings to make changes that will improve overall program quality. They also use this information to create a formal report or description of the program for those outside the program who are interested partners (families, advisory board members, funders, inspection teams, accrediting bodies, etc.). Program leaders will want to use both formative and summative evaluation methods when organizing the program evaluation process.
Formative Evaluation
Formative evaluation is used during the daily operation of the program to examine ongoing processes and to help improve the program. A formative evaluation examines day-to-day successes and deficits. Formative evaluation provides a fast feedback loop to influence program decisions and make necessary changes.
Formative Example: Tori, the T&CS, decides to create a new schedule for herself where she observes two of the toddler classrooms each morning during their snack routine and leaves a brief note summarizing her observation in each teacher’s staff mailbox. Both of the toddler rooms were starting a new routine at snack time and the teachers wanted Tori to observe snack several times to provide feedback about how it was working. At the end of two weeks Tori met with the toddler teachers to discuss her observations and to help them decide if the new routine was working well for the children and staff.
Summative Evaluation
A summative evaluation is typically conducted at the end of a program or after a program has been in existence for some time. The summative evaluation is often shared with those outside the program (advisory board, interagency council, funders) to provide data about the effectiveness of the program.)
Summative Example: Each July, Maria, the Program Manager writes an evaluation report. In it, she summarizes child and youth data, parent event data, child and parent satisfaction data, and the overall program budget report. This report also highlights progress on the program’s goals (e.g., increase the number of volunteers in the after-school program). In many ways, a summative evaluation report may be of interest to anyone affiliated with the program, but is also of importance to individuals outside the program.
Although a summative evaluation report may only be written and shared once per year, the information that is used to create the report is collected across the year at many different points and includes the input of many different stakeholders (e.g., staff, parents, coaches, advisory board members).
Evaluations may be conducted by a staff member who is internal to the program (e.g., T&CS or Program Manager) or the evaluation may be conducted by someone who is external to the program (e.g., paid consultant, university professor with expertise in child care program management, assessor for accreditation, or state licensing representative with expertise in conducting program evaluations). The Program Manager is typically responsible for conducting summative program evaluations, or coordinating and working in collaboration with external program evaluators; however T&CSs may assist at the request of the manager (e.g., to help gather some relevant data).
Planning the Program Evaluation
When planning a program evaluation the main question is "What is the purpose of the evaluation?"
- Do you want to demonstrate why the program needs more child development or school-age program staff?
- Is the purpose to show children's growth and development?
- Is the evaluation being done to gather feedback from others about the program?
The purpose of the evaluation can be different for different stakeholders and will drive the types of questions the evaluation will answer:
- Are we (the leadership team and the staff) doing what we are saying we're doing?
- What is the program's impact on children and their families?
- Do all families feel welcome and part of the program community?
- Does the program comply with DoD certification requirements?
There are typically three major areas of consideration in Child Care and Youth Program Evaluations:
- Program quality
- Program outcomes (for children and youth)
- Consumer and stakeholder satisfaction with the program
It is important that those participating in an evaluation understand the purpose of the evaluation and how their responses will be reported. Planning and designing an evaluation includes creating a matrix in which the major aspects of the evaluation are organized. If the evaluation is not carefully designed then the results may not be accurate or useful to report. The resources in the Apply section will help you more deeply understand the program evaluation process, including what documentation or data may be most appropriate based on the questions you seek to answer, as well as different classroom- or program-quality assessment tools (e.g., environment rating scales) that can offer additional data about the program.
Data used to answer the evaluation questions may be collected through online surveys, focus groups, one-on-one interviews, use of classroom- or program-quality assessment tools, and document reviews (e.g., teacher's written lesson plans, children's progress reports or assessment data, budget summaries, program supply orders).
Model
Sharing and Using Program Evaluation Data
When doing a formative or summative evaluation it is important to plan how the results will be reported and used. Parents will want to know if their time and attention to responding to survey questions results in higher quality child development and school-age programming for their children. Program leaders must decide the best way to share the evaluation results with those who participated in the data collection. Results may be posted on the program's website, in the newsletter, and in other places where families may access it. Hard copies may be made available to staff and advisory board members. It is important for the program manager to also discuss the program evaluation findings with stakeholders. Discussions may take place during a regularly scheduled staff meeting or advisory board meeting. The use of the evaluation results to improve the program should be clear to the stakeholders. Staff and other relevant stakeholders should be made aware about how progress will be monitored.
Some programs are asked to share program evaluation data during interagency council meetings or with other groups. For example, if the school-age program staff and the Program Manager apply for and receive a special grant from an organization (e.g., the Rotary Club) for a substantial amount of money to enhance the science activities offered during the school-age program, then the staff and Program Manager have a responsibility to evaluate the activities and youth learning and report how the money was spent. Again, choosing the most relevant evaluation questions is critical to planning the evaluation, data collection methods, and timeline. The evaluation report should be written, shared and discussed with the funders.
Observe
Watch the video below to hear how T&CSs and Program Managers use program evaluations to document success and guide changes for future improvement.
Model an Approach to Continuous Quality Improvement
The staff looks to the Program Manager and T&CS for leadership. As the leadership team, they focus on a continuous quality improvement outlook. Programs that are committed to consciously improving their services to children, youth, and families demonstrate an openness to learning new knowledge and skills. They have a clear understanding and commitment to quality. They know what a high-quality program should provide for children, youth, and families. They are aware of what the program is doing well and what could be done better. This awareness is based on observation, data collection, and use of valid and reliable program evaluation tools.
There are many different approaches to continuous quality improvement. The below is an example of a process for childcare programs in the state of Rhode Island. (Source: Exceed Rhode Island’s Early Childhood Commitment. (2015). Exceed tip sheet: Continuous Quality Improvement and The Quality Improvement Plan.)
- Assess your Program: What are your program’s strengths and weaknesses? How do you know? In other words, what data do you have about your program, child outcomes, family satisfaction, staff qualifications, etc.?
- Develop and Prioritize Goals: What are your goals for the program?
- How did you arrive at these goals? In other words, what is the current state of the program and how do you know this?
- What tools, documentation or data did you use to understand how your program currently functions?
- Determine Action Steps:
- What are the specific action items to help address these goals?
- What activities or professional development training will help you meet these goals?
- Access Resources: What resource do you need to achieve your goals?
- Complete the Continuous Improvement Plan (CIP):
- Your Service, licensing agency, or accrediting body, may require a specific format for your continuous improvement plan. The Apply section of this lesson provides a sample of the Continuous Quality Improvement Plan (CQIP) for Early Childhood Environment Rating Scale-3 (ECERS-3) from the Illinois State Board of Education and a sample of the CIP used in Ohio. Although each agency, state or Service may be slightly different in how they want the information formatted or delivered, they are typically seeking the same kinds of information:
- Who is responsible for the particular action items listed? For many goals, there may be multiple people responsible, as direct care staff members may be responsible for enacting new practices in their classrooms, while T&CSs are responsible for providing training relative to this new practice and observation and coaching to ensure it is happening appropriately.
- What is the timeline for implementation of the specific action items?
- Evaluate Progress and Revise the Plan as Needed: How will you know your goal was achieved?
You will need to gather information about your progress towards your goal. This may be similar to the information you gathered to identify goals in the first place. For example, if you used the ECERS-3 to set a goal around family engagement, you would also use the ECERS-3 to determine whether you met your goal. Continuous quality improvement entails:
- A commitment to lifelong learning
- Program changes that result in better quality care-giving
- Self-reflection
- Shared leadership
- Embedded job supports
When T&CSs and Program Managers consistently articulate and demonstrate a commitment to continuous quality improvement, they provide a model for all staff to reach for excellence within themselves. The leadership team is always ready to celebrate the program’s successes, but also keeps striving to reach the next goal.
Explore
The families you serve are major stakeholders in any program evaluation. Some programs do written surveys at least once per year (these can be paper surveys or electronic surveys). Some program leaders like to interview parents when their children leave the program (an exit interview) to learn how the program can improve in its mission to create a caring community among the children, staff, and families.
Reflect on the questions that you would include in a family feedback survey. For ideas, review the Sample Family Feedback Survey and Sample Annual School Survey handouts that were adapted from the book Winning Ways for Early Childhood Professionals: Partnering With Families (Schweikert, 2012).
Use the Continuous Quality Improvement-Family Survey activity to practice writing a brief continuous quality improvement plan and think about the goal(s) you would set and the specific action items you would write.
Apply
Many states have a quality rating and improvement system (QRIS) to improve child care services. In some states, a child care’s QRIS rating may be tied to funding, licensing, etc. You may want to do a web search to see if your state has a QRIS rating system, or go to the Child Care Aware State Fact Sheets and look up your state.
Although each agency, state or Service may be slightly different in how they want the information formatted or delivered, they typically seek the same kinds of information. The Illinois State Board of Education’s Continuous Quality Improvement Plan (CQIP) Early Childhood Environment Rating Scale-3 (ECERS-3) and the Ohio Department of Job and Family Services’ Continuous Improvement Plan (CIP) For Step Up to Quality (SUTQ) are two examples of continuous improvement plans.
Glossary
Demonstrate
Carran, D.T. (2009). Early childhood program evaluation. In J. M. Taylor, J. R. McGowan, & T. Linder (Eds.). The Program administrator’s guide to early childhood special education (pp. 307-335). Baltimore: Paul H. Brookes Publishing Co.
Carter, M., Casio, L.M., & Curtis, D. (2020). The visionary director: A handbook for dreaming, organizing, and improvising in your center, 3rd ed. RedLeaf Press.
Center for Early Learning Professionals (n.d.). https://center-elp.org/
Douglass, A., Tout, K., Doyle, S. (2021). Considerations for incorporating the breakthrough series collaborative (BSC) as a quality improvement methodology in early childhood systems. OPRE Report # 2021-211. Office of Planning, Research, and Evaluation, Administration for Children and Families, U.S. Department of Health and Human Services. https://www.acf.hhs.gov/sites/default/files/documents/opre/opre-ccl%20systems%20brief-feb2022.pdf
Exceed Rohde Island’s Early Childhood Commitment. (2015). Exceed tip sheet: Continuous Quality Improvement and the quality improvement plan. https://exceed.ri.gov/Docs/QIP_TipSheet_Feb2015.pdf
ExceleRate Illinois (n.d.). Resources for continuous quality improvement. https://www.excelerateillinoisproviders.com/resources-for-3c
Illinois State Board of Education Early Childhood Division (2017). Program compliance checklist. https://www.isbe.net/Documents/cqip-resources.pdf
Lugo-Gil, J., Sattar, S., Ross, C., Boller, K., Tout, K, & Kirby, G. (2011). The Quality rating and improvement system (QRIS) evaluation toolkit. OPRE Report# 2011-31. Washington, DC: U.S. Department of Health and Human Services, Administration for Children and Families, Office of Planning, Research and Evaluation.
Ohio Department Job and Family Services (2022). Continuous improvement plan (CIP) for step up to quality (SUTQ). https://www.odjfs.state.oh.us/forms/num/JFS01509/pdf/
Schweikert, G. (2012). Winning ways for early childhood professionals: Partnering with families. St. Paul, MN: Redleaf Press.