How to Manage EVS for Optimal Service Quality and Productivity

Transcription

How to Manage EVS for Optimal Service Quality and Productivity
How to Manage EVS for
Optimal Service Quality
and
Productivity
Jo D’Ambrosio, EVS Director
Cottage Hospital, Santa Barbara, CA
Andrew Ganti, MBB
GE Healthcare, San Ramon, CA
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Introduction
Objectives of the Presentation
•
•
•
Define dashboard metrics for balancing quality and productivity in EVS
Select performance targets to promote continuous improvement
Find out how you compare with the best performing peers of yours
Audience
•
EVS Shift & Group Leads, Supervisors, Mgrs., Directors, Hospital CXOs
Contents & Organization
•
•
•
•
•
•
•
Definition of performance monitoring
Definition of a Dashboard and its elements
Need for Balancing the metrics and ways to do it
Some Key metrics in EVS
Examples of Service quality and productivity metrics
Approaches to facilitate continuous improvement
Benchmark to compare the performance with the best in class hospital
Quiz
•
•
•
•
Service Quality:COPIS,Expectations/Metrics, Feedback, Monitoring
Productivity:What is it? Why? How to measure, monitor
Management:What is it? Why? Balance b/w behavioral, financial, Core processes
Optimization:What is it? Why? How to measure, monitor multiple metrics
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Performance Monitoring
• To improve anything, first it should be measured
• Monitoring is to compare expected results with the actual
over time
• Applicable to individuals, groups, sections, departments,
divisions, hospitals, multi-hospitals
• Monitoring may cover one or more aspects e.g. quality,
service, cost, revenue, productivity, profitability, margins
(operating, contributing) etc.
• Monitoring involves metrics or indicators of performance
and can be quantitative or descriptive (e.g. pass/fail,
satisfactory/unsatisfactory)
• Monitoring is not just to measure but to improve, reward
and recognize staff
• Follow-up and feedback are important elements
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
1
Performance Monitoring
Expectations
(standards)
Expectations
(standards)
Svce
Svce
Supplier
Process
Customer
Prod.
Prod.
Feedback
Feedback
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Monitoring Tools
• By User Types - External or Internal
• Dashboards for External Users
• Scorecards for Internal Users
–
–
–
–
–
–
Metrics: single or multiple for balanced
Formats: graphs, tables, gages
Frequency: Daily, Weekly, Monthly ,Yearly
Processing Methods: manual, computerized
Display: Hardcopy or electronic; B&W or color
Dissemination: bulletin board, e-mail, on-line
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
2
Dashboard Components
• Metrics, units & definitions
• Target or Expected performance (can be a
range also a.k.a tolerance limits)
• Actual performance
• Legend, labels & units for axes for a graph
• Interpretation guide I.e. what is favorable
• Variance and Root Cause Analysis
• Corrective Action (what, who, when & how) if
performance is unfavorable
• Feedback mechanism
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
S P E C I F IC W O R K D U T I E S
P O S IT IO N /T IT L E :
Flo o r C a r e
JO B :
3 rd & 6 th F l o o r
E N V IR O N M E N T A L S E R V IC E S
S H IF T :
D ATE:
7 :0 0 - 1 5 :3 0
M ay 2007
E N V IR O N M E N T A L S E R V IC E S
S U P E R V IS O R O N D U T Y
D a ily D u tie s :
•
3 rd
&
6 th E a s t p a t i o n e e d s t o c l e a n d a i l y .
•
S p o t m o p & s w e e p a l l 3 th & 6 th f l o o r H a l l w a y s .
•
S c r u b & b u ff A s s ig n e d r o o m s a s a v a ila b le .
•
C h e c k e le v a to r s la n d in g a n d s c r u b & b u f f a s n e e d e d .
•
C h e ck w aitin g a re a fu rn itu re & cle a n a s n e e d e d .
W e e k ly P r o je c t s :
•
M o n d a y : S c r u b a n d b u f f 6 c e n t r a l h a llw a y s .
•
T uesday : S c r u b an d b u ff 3 c en tr a l h allw a y s.
•
W e d n e sd a y : S c r u b a n d b u f f 3 n o r t h & L . & D .
•
T h u rs d a y : S c r u b a n d b u f f p a t i e n t r o o m s a s a v a i l a b le .
•
F r id a y : S c r u b a n d b u f f 6 c e n t r a l h a l l w a y s .
•
S a t u rd a y : S c r u b a n d b u f f 3 c e n t r a l h a l l w a y s .
•
S u n d a y : S c r u b a n d b u f f p a t ie n t r o o m s a s a v a ila b le .
A d d it io n a l d u t ie s a s a s s ig n e d b y s u p e r v is o r .
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Balancing the Metrics
Types of Metrics
– Behavioral(Customer satisfaction,employee satisfaction, training
effectiveness)
– Financial (dollars related)
– Core Processes (Quality, productivity, response/turnaround
times, discharge cleaning)
Need for Balancing
–
–
–
–
Multiple aspects to any operations
Influencing factors are interdependent
Focusing only on one is undesirable
Reality poses competing priorities
Ways to Balancing
– Prioritize the metrics and select the top few
– Weighted avg. of some metrics if there are too many
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
3
Some Key Metrics in EVS
• Quality (service)
– Cleanliness (per internal/external expectation)
– Response & turnaround time (external expectation)
• Productivity
– By employee, by group, job category
• Operating Cost
– Labor, Supplies & Other
• Satisfaction
– Customer(external/ internal), Staff
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Press Ganey Monthly Results
Dec
Jan
Feb
Mar
Room Cleanliness
80.1%
83.4%
77.9%
78.4%
80.2%
87.3%
Courtsey of Person Cleaning Room
84.9%
85.2%
83.0%
85.0%
85.7%
87.9%
%
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Courtsey of Person Cleaning Room
87
.9
%
%
85
.7
85
.0
83
.0
%
%
84
.9
90%
85
.2
%
Room Cleanliness
100%
April May
80%
70%
87.
3%
80.
2%
78.
4%
30%
77.
9%
40%
80
.1%
50%
83
.4%
60%
20%
10%
0%
Dec
Jan
Feb
Mar
April
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Service Quality
COPIS
– Customer, Outputs, Processes, Inputs, Suppliers
Expectations
– Customers:
• External-Patients & families:
• Internal- Nursing, Ancillary, Support, Medical Staff
Metrics
– Cleanliness by user type (Public, Patient Care, Office
& Other )
– Response time to clean spills to prevent incidents
– Turnaround time to clean and ready the bed after
patient discharge
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
4
Service Quality (Cont’d)
Monitoring
– Data collection
– Data Processing & Dashboard Preparation
– Variance Analysis
– Corrective Action
Dissemination & Feedback
– Dashboard data display, recognition
– Variance explanation & follow-up support
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
DIRTY TO CLEAN BED TRACKING
WK 1 WK 2 WK 3 WK 4 WK 5 WK 6 WK 7 WK 8 WK 9 WK 10
DIRTY
CLEAN NEXT
STAT
Total Number of
Rooms Cleaned
64
66
65
66
64
59
64
65
65
63
43
47
41
49
38
44
41
41
46
41
36
35
40
35
33
27
37
35
32
39
789
779
749
772
809
658
741
744
821
847
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Productivity
COPIS
–
–
–
–
–
Customer (See below)
Outputs (Dashboard, Var. Explanation, Corr.Actions)
Processes (collect data, prep dashboard, distribute)
Inputs (Time, $, Sq.ft., count of other units of measure)
Suppliers (Finance, EVS)
Expectations
– Customers: Internal- CXO, Multi-Dept.Director, Mgr., Sup., Lead
Definitions:
– Traditional: Productivity = Output/Input e.g. Meals/
hour;
– Modified: Productivity = Input/Output e.g. hours/meal;
Nsg hours/pat. day
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
5
7/2/2007
Reports for BedTracking® Statistics
Santa Barbara Cottage Hospital
0
Response/Turn-Time
Date: 6/29/2007 , All Users , All Patient Units , Cleaning time 5 - 60 min. Max. Response
time 120
min. , Based on Last Requests
ID:1
Employee 1
Status
Turn
Requests Compliant/Adjusted
Avg. Response Avg. Clean
Avg.
Cleans
Dirty
13
8 (62%) / 5 (38%)
1:00
0:19
1:19
Clean next
4
4 (100%) / 0 (0%)
0:10
0:22
0:32
Stat
2
2 (100%) / 0 (0%)
0:05
0:13
0:18
Totals (all statuses)
Requested: 19
Compliant Cleans: 14 (74%)
Adjusted Clean: 5 (26%)
ID:0
Employee 2
Status
Turn
Requests Compliant/Adjusted
Avg. Response Avg. Clean
Avg.
Cleans
Dirty
7
6 (86%) / 1 (14%)
0:44
0:16
1:00
Stat
1
1 (100%) / 0 (0%)
0:16
0:23
0:39
Totals (all statuses)
Requested: 8
Compliant Cleans: 7 (88%)
Adjusted Clean: 1 (13%)
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Metrics:
Productivity (Cont’d)
Time per unit of measure or Units (of measure) per time
unit
Time = Worked or Paid; Hours or Minutes
Unit of Measure:
– Area cleaned in a day, pay period, month or year per 1000 sft;
gross or net; sft by area such as public, Pt. care by type:
Med/surg, ICU; routine vs. isolation Office,hallways, stairs;
– Units cleaned in count I.e. No. of bathrooms cleaned; patient
days, adjusted patient days
– e.g. Worked hours per pay period per 1000 net sft.
Cleaned; per Adj. Pt. Day; per Adj. Discharge
Conversion factors: Paid to worked, divide by 0.9
Hrs. per year to Mo., divide by 12, Hrs/mo. to Hrs/pay period,
multiply by 12 and divide then divide by 26
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
_______________________________ Hospital
Metrics Development Worksheet
Performance Measurement: proposed
Proposed Measurement: ____________________________________________
How will the data be collected? _______________________________________
How will the proposed measurement be calculated? _______________________
What will be included or excluded in the measure? ________________________
With what frequency will the indicator be reported (monthly, quarterly) ________
Goals/Benchmarks:
What is the goal for the measurement? _________________________
What is the source of the measurement goal? ___________________________
Performance Improvement:
How does this indicator relate to the organizational goal of Patient Safety?
Signature ______________________________
Date ________________
Approval _________________________
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
6
Productivity (Cont’d)
Monitoring
– Data collection (Type, duration, sample size)
– Data Processing & Dashboard Preparation
– Variance Analysis
– Corrective Action
Dissemination & Feedback
– Dashboard data display, recognition
– Variance explanation & follow-up support
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Management
Definition:
–
–
–
–
–
Doing more with Less;
Doing the right thing at the right time;
Processing by priorities
Balancing competing priorities
Balancing competing forces e.g. staff needs vs. dept.
needs in staff scheduling
Purpose: Balance b/w behavioral, financial & core processes
Tools: Dashboards, Standard Operating Procedures
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Optimization
Definition:
– Maximum value obtainable when there is more than
one variable e.g. Optimum speed to get the highest
gas mileage in miles per gal. is 55 mph.Any speed
above or below 55 reduces the gas mileage
– Quality vs. Productivity i.e. productivity can be
increased up to a point after which quality could suffer
– Measure and monitor more than one metric that is
likely to move in the opposite direction
Monitoring:
– Watch for the trends in two or more metrics to ensure
one is not improved at the expense of the other
– Note that improving multiple metrics is not mutually
exclusive but has an optimum level
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
7
Approaches to Continuous Improvement
Definition: Improvement is a journey and not a destination. That is
why the improvement needs to be continuous
Approaches/ Methodologies:
• CQI: Continuous Quality Improvement
• TQI: Total Quality Improvement
• SDCA: Standardize (stabilize), Do, Check, Act
• PDCA: Plan, Do, Check and Act
• Six Sigma: DMAIC (Define, Measure, Analyze, Improve, Control)
to go to 6 sigma level or almost zero(99.999 good) defects
• Lean: Toyota Production System using Value Stream Maps, Single
Piece Flow, Just in Time, Line Balancing and 5 S (Sort, Sweep,
Shine, Standardize & Sustain) as part of Kaizen bursts
• Benchmarking & Best Practices with Peer or Best in Class outside
the hospital industry comparisons
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Benchmarking & Best Practices
Definition: Improvement is a journey with multiple milestones.
Benchmarking with peers is one way to select the target or the next
milestone and the best practices help reach that in the continuous
improvement process. Each can be used independent of the other
Benchmarking in EVS:
Metrics with benchmark data:
– Service Quality: Cleanliness, Response & Turnaround times
– Productivity: worked hrs/1000 sft, Worked hrs/adj. Discharge,
Worked hrs/patient day; worked hrs/discharge
Benchmarking Data Sources:
–
–
–
–
–
Hospitals in the multi-hospital system
Informal networking
Local, state and national trade association(?)
Contract Mgmt. Co.;EVS equipment/supplies vendors
Commercial benchmarking co. (e.g. Solucient)
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Benchmarking Metrics - Examples
SKILL MIX
Custodial Worker %
Management %
Other %
Hours Worked: Hours Paid %
Non-Payroll Hrs: Total Hrs Paid %
WORKLOAD / SERVICE INTENSITY
Non-Hospital Sq Ft: Tot Net Sq Ft %
Disch / Trans Cleanings / Discharge
LABOR PRODUCTIVITY RATIOS
HOURS WORKED PER
1000 Net Sq Ft Cleaned/Pay Period
Adj Patient Day
Adj Discharge
COST RATIOS
Staff Average Rate / Hour
Regional Adj Rate / Hour
EXPENSE $ / 1000 NET SQ FT CLEANED
Labor
Regional Adj Labor
Other Direct
Cleaning Supply
Management Contract
Total
Total w. Regional Adj Labor
EXPENSE $ / ADJ PATIENT DAY
Labor
Regional Adj Labor
Other Direct
Cleaning Supply
Management Contract
Total
Total w. Regional Adj Labor
EXPENSE $ / ADJ DISCHARGE
Labor
Regional Adj Labor
Other Direct
Cleaning Supply
Management Contract
Total
Total w. Regional Adj Labor
Total Dept Exp: Total Hospital Exp %
Source: Solucient
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
8
Adjust for the Differences
Statistical Characteristics
– % Carpeted, % Infectious wasted, % Spl proj. etc.
Hours include Mgr. or not
Organization Characteristics
– Clean rooms after discharge/transfer,Setup mtg.
Rooms,Clean kitchen floors, grounds, Replace light
bulbs, move beds. Furniture etc.
Dept. Operational Characteristics
– Staff assigned to OR, OB, ED; Contract mgmt.,
Computer system, special equipment., floor storage,
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
Quiz
• Service Quality
– COPIS
– Expectations/Metrics, Feedback & Monitoring
• Productivity
– What is it? Why? How to measure it? Monitor it?
• Management
– What is it? Why? Balance between behavioral,
financial, core processes, example for each metric
• Optimization
– What is it? Why? How to measure & monitor multiple
metrics
2007 ASHES Annual Conference
September 30-October 4
St. Louis, MO
9