Previous Page  6 / 32 Next Page
Information
Show Menu
Previous Page 6 / 32 Next Page
Page Background

6

/ Canadian Government Executive

// December 2016

Deliverology

abound, as do reports of its overall effec-

tiveness. Our research categorizes the les-

sons learned as follows:

Practical Considerations for

the Transition to the new

Policy on Results

Transitioning to the new Policy on Re-

sults is not just a paper exercise intended

for the Planning and Performance Mea-

surement specialists in Corporate Ser-

vices. Every Program Owner (or Program

Official) needs to be directly engaged in

making clear the results that their pro-

gram contributes to their department’s

priorities and mandate. The linkage back

to the Minister’s mandate letter is imper-

ative.

In addition, other disciplines such as

policy, finance, data management and oth-

ers should be involved to understand the

implications of the new Policy on Results

constructs and to ensure broader buy-

in. Chief Results and Delivery Officers

(CRDO) and Chief Data Officers (CDO)

are in the process of being put in place;

but theirs is a coordinating role, not a dic-

tating one.

Some departments are simply port-

ing their Program Activity Architecture

(PAA) content into the new Departmental

Results Framework (DRF), while others

are taking a “clean sheet” approach to re-

thinking how programs are defined and

structured.

Some are going further – updating their

Logic Models or Outcomes Maps to en-

sure their programs can tell a comprehen-

sive “performance story”. This will then

drive which Key Performance Indicators

(KPIs) are required, and not the other way

around. Then the whole conversation

about targets, results profiles and toler-

ances can begin – with the bigger picture

of program performance in mind.

Finally, although it has likely been said

of every significant change across an or-

ganization, sustained senior management

commitment is critical. It will affect the

programs they oversee, their accountabili-

ties and ultimately their department’s suc-

cess. And that is exactly what Deliverology

is meant to achieve.

G

regory

R

ichards

is the Director of

the MBA Program and of the Centre

for Business Analytics and Performance

at the Telfer School of Management at

the University of Ottawa. Richards@

telfer.uottawa.ca

C

arrie

G

allo

is a Partner in the

Advisory Services Practice with Interis

| BDO.

cgallo@bdo.ca

M

urray

K

ronick

is a Senior Manager in

the Planning and Performance Practice

with Interis | BDO.

mkronick@bdo.ca

Table 1.

Summary of Lessons Learned from Deliverology Implementations

Lesson Learned

Comments

Sustained senior management

and political interest in the

policy outcomes are critical

for success.

The implementation will not

survive without credible data.

It is important to create docu-

mented routines for using the

data to drive program changes.

Measurement can create a

tendency to manage towards

the performance standard.

Silo effects can occur.

It’s important to align the

organization’s “philosophy

of management” with the

Deliverology concept.

It’s a truism to claim that nothing succeeds without senior leadership interest. In many

cases, “interest” is often demonstrated by signing off on the project proposal. In the

successful cases of Deliverology implementations, senior leadership involvement went

much further. Politicians and senior departmental leaders were involved in reviewing regular

progress reports and contributed to helping remove barriers to success where needed.

So, “interest” is not simply tacit agreement that the outcomes are important, it also includes

direct, visible and sustained involvement to help remove barriers and allocate resources

where needed.

“Objectively verifiable” data is critical. This avoids arguments about the facts so the Delivery

Unit and program managers can get on with problem resolution / mid-course correction.

Successful implementations had a plan of some sort for what happened to the data and the

analysis: who gets it, who has decision rights, and how to evaluate the impact of changes.

A great deal of the criticism of Deliverology revolves around “gaming” of the system to

meet the established performance standard.

For example, in the education sector, teaching children only what they need to know to

pass standardized tests obtains the result being measured, not necessarily the broader

outcome of knowledgeable students.

Silo effects occur when certain groups are identified for additional attention, based on

pre-established performance targets; the others can, by default, be identified for neglect.

Sticking with the education sector as an example, some children performing far below the

performance standard would be selected for extra help. The others at the standard could

be ignored until everyone has caught up.

Deliverology has also been criticized for its apparent “command and control” culture. There

are different ways of implementing it, however, and highly successful implementations have

been able to push decision-making down to the front lines assuring that people who are

doing the work are empowered to improve it.

Some level of control is always needed, but it is important to match the control mechanisms

with the mandate and the risk profile of the organization.