Matorera Management 2016
Matorera Management 2016
Matorera Management 2016
Douglas Matorera
In
Faculty of Education
Signature
Douglas Matorera
i
Abstract
At attainment of independence in 1980 the Zimbabwe Government adopted an egalitarian
approach to the provision of education. Primary education was made compulsory and by
2008, the number of primary schools had increased by 136.98% and enrolments by
1980.38%. At the secondary school level the number of schools increased by 1132.77% with
enrolment galloping to a figure of 1155.74%. At pre-university level student enrolments shot
to 1125.86%. The ripple effect propagated right into higher education, putting constraints on
quality of academic staff, quality and supply of resources, and the content and processes of
management, leadership and instruction. A gradual reduction in State funding and its total
withdrawal for postgraduate level forced universities to device own self-funding strategies.
To meet the multiplicity of demands from multiple stakeholders the Chinhoyi University of
Technology Graduate Business School (CUTGBS) adopted a QFD approach.
The purpose of the case study was to assess and evaluate the response of a university
business school (CUTGBS) to a QFD-based model for assuring quality in a structured
master’s degree programme. A multi-method approach that included in-depth interviews,
focus groups, documents analyses and observations was used to assess and evaluate the
response of staff in the CUTGBS to the model and to the manner in which the model was
rolled out. Interviews involved academics, non-academic staff, students, alumni and senior
managers within the CUTGBS. Interviews were also held with staff in the Ministry of Higher
Education and Training, industry, and the Zimbabwe Council on Higher Education
(ZimCHE), the national quality assurance agency.
The purpose of the study was to contribute to our understanding of how staff in the
educational services sector respond to Business-based quality models by assessing and
evaluating the adoption of QFD in a university Business School. A better understanding of
quality management in terms of the tools and stages of the QFD model should create a new
dimension of quality management in higher education against the domains of context, inputs,
processes and market-orientation.
ii
The CUTGBS took a structured and deliberative approach in the adoption of QFD. A core-
team of permanently employed academics and non-pedagogic staff constituted the core QFD
team which had the key strategic role for the CUTGBS. There were evidences of team-work-
quality on the parameters proclaimed by Hoegl and Gemeunden (2001). The CUTGBS staff
participated in running ‘Voice of Customer’ based on a Six Sigma paradigm. There was
marginal use of Affinity Diagram, Tree Diagram and Kano’s model in the treatment of Voice
of Customer, a situation that has the danger of skewing decisions in favour of the
domineering members of the team. However in situations of ‘ideal speech situation’
(Harbermas, 1995) and deliberative democracy (Gutmann and Thompson, 2004) it has the
advantage of deep-going aspect-by-aspect deliberations. Much of the normative and
regulative requirements of the quality assurer were treated as ‘voice of market’ and thus
escalated to the CUTGBS policy regime.
Student Evaluation of Teaching Effectiveness (SETE) were used for Customer Satisfaction
Performance evaluations, Goal Setting, and for scoping Target Value within the QFD
paradigm. In essence QFD was adopted creatively and implemented on a selective
incremental approach.
Being a case study, the findings of this study have an indicative rather than a conclusive
value. However the validation study indicated the feasibility of using QFD as a quality
assurance model within the higher education system.
Key words
Quality, deployment, function, constraint, strategy, noise, categorisation, interface-mapping,
case-owner; target-value
iii
Acknowledgements
My acknowledgements go to the staff at the Chinhoyi University of Technology, the
Chinhoyi University of Technology Graduate Business School, the Zimbabwe Ministry of
Higher Education and Training, the Zimbabwe Council on Higher Education who took their
time and efforts to participate in this research. I felt very comfortable around them.
I should thank Professor W. J. Fraser for his guidance without which this study could have
been extremely difficulty.
My special thanks to everyone who gave time for the discussions during the validation phase
of the study findings. I should thank everyone who in whatever way made a positive
contribution to this study.
iv
TABLE OF CONTENTS
v
2.3.6.1 Gaining alignment 28
2.3.6.2 Embracing change 29
2.3.6.3 Establishing a learning culture 30
2.3.6.4 Relating the micro (individual) to the macro (company-wide) 31
2.3.6.5 Measuring and reporting 32
2.3.6.6 Supporting distributed leadership 33
2.3.6.7 Being up front 34
2.3.7 Strategic risk management 35
2.3.7.1 Resourcing for the medium term 35
2.4 QFD: The three perspectives 38
2.4.1 The quality perspective 38
2.4.2 The function perspective 44
2.4.3 The deployment perspective 45
2.5 Early uses of QFD 45
2.6 What QFD is being used for today 46
2.7 Customer: The focus of a QFD strategy 51
2.8 The growth and philosophy of QFD 51
2.9 QFD and the quality-innovation helix 53
2.10 QFD as a high-level strategy planning model 54
2.11 Application of QFD in Education 55
2.12 Stumbling blocks for implementation of QFD 61
2.13 Conclusion 62
vi
3.4.2 The behavioural objective model 74
3.4.3 The goal-free evaluation model 75
3.4.4 The success-case model 75
3.4.5 The utilisation-focused evaluation model 76
3.4.6 The empowerment evaluation model 76
3.4.7 The collaborative model 76
3.4.8 The organisational learning model 77
3.4.9 The consumer-oriented approach 78
3.4.10 The responsive evaluation approach 78
3.5 Organisational culture and model institutionalisation 79
3.5.1 Introduction 79
3.5.2 Level of institutionalisation 79
3.5.3 Managing the link between quality model (strategy) and 81
organisational culture
3.6 Inconsistencies in quality assurance 83
3.6.1 Introduction 83
3.6.2 Disconnect between purpose and metrics 84
3.6.3 Accountability as a mode of quality assurance 86
3.7 Dimensions of quality assurance 88
3. 8 Global and international efforts at quality assurance 90
3.9 Stakeholders’ views of the purpose of higher education 91
3.10 Philosophy and methodology of quality assurance and the idea of an 93
ideal master’s degree graduate
3.11 Improving the quality of quality assurance work in higher education 95
institutions
3.12 Features of an effective quality assurance system 96
3.12.1 Impersonal attributes 96
3.12.2 Human competences 97
3.13 External Quality Assurance mechanisms 98
3.13.1 Introduction 98
3.13.2 Licensing 99
vii
3.13.3 Accreditation 99
3.13.3.1 Introduction 99
3.13.3.2 Expectations on the accrediting process 100
3.13.3.3 Objectives of accreditation 100
3.13.3.4 Benefits of accreditation 102
3.14 Professional associations and their influence on quality assurance 103
3.15 Impact of external examination on quality assurance 104
3.15.1 Introduction 104
3.15.2 Work of external examiners in determining quality 104
3.15.3 Difficulties with external examination 105
3.15.4 Improving external examining 106
3.15.5 Rankings and their impact on quality assurance 107
3.16 Modelling programme quality assurance through QFD: aligning and 107
integration
3.16.1 Introduction 107
3.16.2 Context Analysis and Static-Dynamic Status Analysis 108
3.16.3 Structuring of product design specifications 109
3.16.4 Pugh Concept Selection 109
3.16.5 Total system / subsystem analysis 109
3.17 QFD’s House of Quality 110
3.17.1 An overview 110
3.17.2 Strategic planning 114
3.17.3 Establishing the QFD team 117
3.17.4 Voice of Customer 120
3.17.4.1 Introduction 120
3.17.4.2 Deficiencies of Voice of the Customer tools 121
3.17.4.3 Variations in meaning of Voice of the Customer 122
3.17.4.4 Escalating Voice of the Customer into Six Sigma roadmaps 123
3.17.5 Regulatory Requirements 134
3.17.6 The Product Planning Matrix 137
3.17.7 Competitive Satisfaction Performance 139
viii
3.17.8 Technical Descriptors 142
3.17.9 Goal Setting and Improvement Ratio 142
3.17.10 Relationship Matrix 146
3.17.11 Institutional Difficulty / self – assessment 148
3.17.12 Technical analysis of competitor products 150
3.17.13 Target values for technical descriptors 151
3.17.14 Correlation Matrix 153
3.17.15 Absolute importance 154
3.18 QFD and other strategy models 156
3.19 Conclusion 160
ix
4.10 Conclusion 188
x
5.7 Staff perception on the implementation and institutionalisation of 242
the QFD model for the purposes of quality assurance in the M.Sc.
Programme.
5.7.1 Introduction 242
5.7.2 Response of the CUTGBS to internal quality assurance 243
5.7.3 CUTGBS’s response to external quality assurance mechanisms 247
5.8 Management’s response to the results of the application of the QFD
model 248
5.9 Conclusion 251
xi
References 278
LIST OF FIGURES
Fig. 1.1 The three aims of the study as synthesised from the literature 5
Fig. 2.1 Evolution of the idea of quality assurance 13
Fig. 2.2 Modified Strategy Focus Wheel 17
Fig. 2.3 The relationship between ten determinants of quality, perceived and
expected quality of a service 39
Fig. 2.4 A QFD-based model for insuring deployment of quality concepts as
synthesised from 42
Fig. 2.5 Kano’s model 56
Fig. 3.1 Using QFD to meet PQA in terms of needs of the various customers 73
Fig. 3.2 The looped incremental relationship between elements of
institutional work 80
Fig. 3.3 Accountability for, to, and how 87
Fig. 3.4 Two dimensions to quality assurance 89
Fig. 3.5 The traditional epistemology of quality education 93
Fig. 3.6 Four-phase House of Quality 111
Fig. 3.7 Six-roomed House of Quality 112
Fig. 3.8 Thirteen stages of the QFD’s House of Quality 115
Fig. 3.9 Six Sigma approach to Voice of Customer 123
Fig. 3.10 Six Sigma roadmaps linked to Voice of Customer 127
Fig. 3.11 Voice of Customer and Marketing in customer-orientation 128
Fig. 3.12 The Design Scorecard and elements of Product Planning 139
Fig. 3.13 Correlation matrix building into the House of Quality 154
Fig. 4.1 Methodological triangulation 169
Fig. 4.2 Approach to thick description and validity of the research 170
xii
Fig. 4.3 Alignment between the research question, data sources and evidence
generated 178
Fig. 4.4 Relationship between number of interviewees and cumulative 180
percentage of valuable data
Fig. 4.5 Movements between a transcribed interview and other data 183
Fig. 5.1 Communication matrix and the dilemma of how-much? 192
Fig. 5.2 Definitional aspects/characteristics of QFD 195
Fig. 5.3 The dual nature of QFD and its relation to market-orientation 197
Fig. 5.4 Organisational challenges faced by the CUTGBS 201
Fig. 5.5 Stages of the QFD model 206
Fig. 5.6 CUTGBS’s mode of adoption of QFD 236
Fig. 6.1 The four senses of my research problem 253
Fig. 6.2 The relational nature of context, model, strategy, actions and 258
outcomes
Fig. 6.3 QFD influencing the way strategy planning, change-project 259
management , implementation and risk management were done
Fig. 6.4 A holistic construct of quality as synthesised from literature and 262
Fig. 6.5 The different demands and their influence on graduate attributes as 269
synthesised from literature
Fig. 6.6 The relationship between purpose of education and philosophy of 270
achieving that purpose
Fig. 6.7 An approach for linking model to context, strategy and desired 274
outcomes
Fig. 6.8 ‘Designing down’ from ideal graduate characteristics to the course 275
LIST OF TABLES
Table 2.1 Relation between Strategic focusing and the Best-practice principles. 22
Table 3.1 The many hats of the strategist / manager 119
Table 3.2 DFSS - Six Sigma approach to Voice of Customer 126
Table 3.3 SWOT Analysis 149
Table 3.4 Comparison between BPR, TQM [QFD] and Six Sigma 158
xiii
Table 4.1 Research design 172
Table 5.1 Enhanced SWOT Analysis 233
LIST OF APPENDICES
Appendix 1 Definitions of constructs used in portraying a meaning of QFD 313
Appendix 2 Interview schedule for staff on strategy for quality in the CUTGBS 315
Appendix 3 Structure of the M.Sc. Programme 316
Appendix 4 Interview questions for staff on implementation of QFD stages 319
Appendix 5 Strategic tools used within QFD (philosophical orientation) 323
Appendix 6 QFD tools used within QFD’s methodological orientation 324
Appendix 7 Questions on the link between QFD and external quality mechanisms 326
Appendix 8 Interview schedule for staff on management’s response to results of 327
QFD
Appendix 9 Interview schedule for the ZimCHE Directors 328
Appendix 10 Interview schedule for students 329
Appendix 11 Observation schedule: lecture halls 330
Appendix 12 Observation schedule: library 331
Appendix 13 Schedule for document analysis 332
Appendix 14 Questions for the validation study 333
Appendix 15 Letter of consent to participate in the research 335
Appendix 16 Chinhoyi University of Technology Research approval letter 336
Appendix 17 University of Pretoria Ethics clearance certificate 337
Appendix 18 University of Pretoria Integrated declaration form 338
LIST OF ACRONYMS
AACSB Association to Advance Collegiate Schools of Business
ABET Accreditation Board for Engineering and Technology
APQN Asia-Pacific Quality Network
CUTGBS Chinhoyi University of Technology Graduate Business School
DFSS Design for Six Sigma
xiv
EQUIS European Quality Improvement System
INQAAHE International Network for Quality Assurance Agencies in Higher
Education
IUC-EA Inter-University Council of East Africa
KASUB Knowledge Attitudes Skills Understanding Behaviour
MFSS Marketing for Six Sigma
ODT Optimal Distinctiveness Theory
PQA Programme Quality Assurance
QA Quality Assurance
QAA Quality Assurance Agency
QFD Quality Function Deployment
SSPD Six Sigma Process Design
SBT Strategic Balance Theory
TFSS Technology for Six Sigma
LIST OF ABBREVIATIONS
M.Sc. Master of Science
xv
CHAPTER 1 – PROBLEM STATEMENT, RESEARCH RATIONALE AND RESEARCH
QUESTIONS
1.1 Introduction
This chapter introduces the research in terms of its context and outlines the structure of the
thesis document. The next section is a highlight of the post-independence expansion of the
education system, which resulted in institutional and programme massification towards the
end of the 1990s. A statement of the research problem is presented in terms of its four senses
– source, primary source, context and action senses. The aims, rationale, research objectives
and the research questions of the study are discussed. The chapter closes with a presentation
of the structure of the thesis document.
By 2013, there were 15 operational universities in Zimbabwe, nine of which were state
universities and the other six being private universities. The total number of state universities
was projected to go up to 12 by 2015, bringing the total number of universities to 18 for a
population slightly under 13 million. There were also three polytechnic colleges that were
teaching undergraduate courses under the mentorship of senior universities: National
1
University of Science and Technology, University of Zimbabwe, and Midlands State
University. There are currently more than ten polytechnic and 30 normal colleges that are
affiliated to different universities, with their compound number of programmes running well
into thousands.
Internationally, the accelerated increase in undergraduate output has been reported as having
a ripple effect of creating an enormous demand for postgraduate education and triggering the
proliferation of demand-absorbing (Lin et al, 2005; Bing, 2009; Wu, 2009) diploma and
degree mills (Altbach et al., 2009:xii), garage universities (World Bank Report, 2000:32),
and pseudo universities (Altbach et al., 2001:8) offering low-quality programmes (Lomas,
2002:1; Gamage, et al., 2008; Altbach, et al., 2009; Gregorutti, 2011). The Organisation for
Economic Cooperation and Development (OECD) (2008) observed the doubling of higher
education student population from 68 million in 1991 to 132 million by 2004 and today
higher education qualifications are ubiquitous. In his prize-winning paper, Akerlof (1970)
described poor-quality goods as ‘lemons’ and warned that they have the danger of replacing
products and services of higher quality and of greater use to buyers and users. This amounts
to low-quality, less rigorous, less relevant and unfit-for-any-purpose programmes replacing
more rigorous, more relevant, value-adding and fit-for-purpose programmes. The long-
standing debate on the quality of education appears to have gone thicker, with more voices
and greater and more pronounced demands that HEIs must show what they are doing about
improving the quality of their institutions, curriculum, graduates and everything in between
(Shah and Nair, 2014:148; El-Khawas, 2014:183). The responses have been multiple in form
and in content, with some HEIs adopting new structures, models, policies and strategies (El-
Khawas, 2014:184). The results range from disappointments to something to write home
about and to celebrate. Explanations about either success or failure are equally numerous.
These forces are shaping higher education discourse and practice across the world (Krucken,
2011:1). In the fast transforming higher education landscape most business schools have
been pressured to adopt business strategies in the hope that they would be more competitive
by increasing stakeholder satisfaction (Crebert, 2000:74; Abdous, 2011; Zineldin and
Vasicheva, 2012:65; Caspersen et al., 2014:195). The number of higher education institutions
that have transcended from traditional bureaucratic management styles to adopt variants of
the New Public Management (NPM) and managerialism (Shore and Wright, 2000; Strathern,
2000; Morley, 2003; Blackmore, 2009) has been increasing over the years. In spite of this
increase, research publications have remained ambivalent on whether the relevance of
education has improved over the years as a result of the adoptions (Nusche, 2008; Remler
and Pema, 2009; Haggis, 2009; Coates, 2009 / 2010; and Trigwell, 2011). In actual fact, what
has attracted some inquiry is the compatibility of the industry-based model with the
education sector; whether management in higher education institutions would be willing and
able to contextualise the models; how to mediate the differences between the industry-based
nature of the models and the services sector-based environment; and dealing with project
change management that is called for by most adoptions (Kohn, 1993; Owlia and Aspinwall,
1997:279; 541; Franceschini, 2002:117; Pompili, 2010:239; and Narang, 2012:359).
In Africa, such models have been applied in Ethiopia, Nigeria and presumably other
countries (Mohamedbhai, 2008). While some researches assert that some of the successes of
3
higher education institutions are anecdotal, many authors feel that management should
shoulder the blame for failure rather than the models they use (Keller, 1993; Ho and Wearn,
1995; Birnbaum and Deshotels, 1999; Crebert, 2000; Franceschini, 2002; Senge, et al., 2007;
Stensaker, 2008; Chan, 2009; Ficalora and Cohen, 2009; Mukaddes et al., 2010; Bolden, et
al., 2010).
QFD is not widely understood, let alone tested in higher education in its completeness.
Driving it to the central core of institutional mental model and the institution’s fabric of
behaviour is not supported by dedicated theory and we live by the assumptions that it should
call for sets of skills which need time to build and perfect. The adoption and roll-out of QFD
should be further complicated by its multiple-stage nature and how people come to share
visions around its governance, diffusion and strategic value in the contexts of previously
established models of management and services delivery. Many find QFD tools numerous,
too technical in their nature and unfamiliar to the discipline of education yet they find the
philosophy of QFD beguiling.
In 2005 the Graduate Business School (CUTGBS) of the Chinhoyi University of Technology
(Zimbabwe) adopted a QFD-based model as a strategic framework to guide its efforts of
developing a high quality M.Sc. Programme in Strategic Management. It is in this backdrop
that the research problem was to contribute to our appreciation of quality and quality
assurance and to the growing but still limited understanding of the adoption and
implementation of New Public Management (business) models by examining how the
CUTGBS adopted and diffused the various stages and tools of QFD and its implication on
the quality of the M.Sc. Programme. Plugging research findings with literature on QFD and
models traditionally applied in higher education, theories of organisation management,
education and institutionalisation should help us expand our understanding of the dynamics
of quality management particularly using business models.
5
1.5 Research objectives
The study was driven by five main objectives:
To explain the characteristics and nature of QFD
To discover the motivation of the CUTGBS in choosing the QFD model
To assess staff response to the QFD model and the way it was being institutionalised
To evaluate the intensity of use of QFD tools and techniques and the implication on the
level of adoption of QFD
To discuss staff perceptions on the implementation and institutionalisation of the QFD
model for purposes of quality assurance in the M.Sc. Programme.
To evaluate the responses of management to results of application of the QFD model.
1.9 Conclusion
8
CHAPTER 2 – THEORETICAL FRAMEWORK, CONCEPTUAL FRAMEWORK AND
A DESCRIPTION OF QUALITY FUNCTION DEPLOYMENT
2.1 Introduction
This chapter discusses the theoretical and conceptual frameworks undergirding the study. It
also discusses the QFD model. In section 2.2 the theoretical perspective for this study is
explicated. The set of assumptions underlying the research is that the success or failure of an
organisation reflects the ability of those in it to manage the organisation’s risk
interdependency field. In section 2.3 the conceptual framework is discussed. A five-strand
Strategy Focus Wheel, with each strand supported by some management best practice
principle(s) is discussed. These best practice, also called excellence principles are the vehicle
and media of deployment of quality throughout the organisation’s structures, functions,
interfaces, processes and the mental models that design and put the principles to work.
In section 2.4 the three perspectives of QFD are presented and I argue that it matters a lot to
be unambiguous about ‘quality’, about how to pursue it and about the most favourable
organisational culture. From section 2.5 through 2.8 I discuss the use, growth and philosophy
of QFD. By means of a handful literature reviews (section 2.9) I show that the use of QFD in
education is no longer numinous. In section 2.9 the quality-innovation helix is presented. The
core of the helix is that in QFD quality is a practice that hinges on protracted improvements
and innovation and not on inspections and compliances designed by a few. This perspective
calls for a huge change in the ways the professoriate should carve their place and define their
roles within the totality of university actorhood. See next section for an explication on
actorhood. Section 2.10 presents QFD as a high-level strategy planning tool with 2.11
discussing how this role can be constrained by poor relationships, policy, resources and their
interfaces. This point is reiterated in the Theory of Constraint which is the theoretical
perspective underpinning this study. .
Achieving the most effective and efficient way of transforming resources into goals is also
the sublime aim of QFD. So why deploy the theory of constraint when you are already using
QFD? In the early stages of the adoption of QFD, it is important to have a thorough strategic
plan in place to identify risks / constraints and to detail and profile them. Such action helps us
to understand the organisation’s risk interdependence field and its mobility. QFD offers the
methodology of embedding the risk-response strategies into the processes of the organisation.
The basic assumptions of the Theory of Constraint also apply to QFD, Six Sigma and other
management models, and include the following:
The view of an organisation as a mobile network of interdependent structures, functions
and processes (Moscovic, 1998; Elias, 2000; Flick and Foster, 2007; and Flick, 2007:19).
Elements of the phenomenological figuration perspective are adopted in this study.
The view that there is an interdependence relationship between policy, structure-
structure; structure-function and function-function relations and their interfaces and that a
mosaic of cause-effect relations exist among these organisational factors (Pearce and
Robinson, 2009).
The sublime aim of organisational management is to optimise organisational performance
and be out front (Bryson and Alston, 2005:47; Bevington and Samson, 2012:49). The
study assumes that the adoption of QFD in the CUTGBS was done with some conscious
goal relating to enhanced quality performance in the minds of management and staff.
Each organisation has at least one major constraint and ignoring or disregarding the
distinction between constraint factors and non-constraint factors inevitably leads to
wrong decisions. When a model is adopted it is important to be sure that no constraints
are brought in with the model apart from the many others that may already exist or will
arise due to the numerous adoption activities (Oakland et al., 2002:1126).
10
High-performance organisations are fundamentally ‘systems’ genomes that leverage
performance on a 360-degree vision of everything and they consider everyone as equally
important (Dougherty, Barnard and Dunne, 2005:38; Bevington and Samson, 2012:172).
Different organisations experience different risk interdependency fields characterised by
multiple and complex cause-effect relationships (Roberts et al., 2005:9).
Humans think and act according to how they live and shaping relations and perceptions
and garnering stakeholder engagement is a critical success factor for any strategy and its
implementation (Jauch and Orwig, 1997:280; King, 2004:11; Temtime and Mmereki,
2010; Jones, 2014:13).
The view of the Theory of Constraint held in this study is that an organisation like the
CUTGBS is a purpose-oriented and ever-evolving organic entity constituted by a plethora of
factors that form a web of interdependent relationships in which a change with one factor
ripples some modification throughout the web of the relationships. This new view of the
higher education institution, like the CUTGBS as an actor (Steenkamp et al 2012:380)
alongside other actor-institutions in society pervades this study. This view fits in with the
three fundamental sociological perspectives: structural-functional, symbolic interactionism,
and social conflict. Conjointly they view an organisation as multiple complex structures, each
with sets of functions and the inequality among the structures and the diversity of functions
continually generating conflict and change. Any such change is likely to improve or constrain
performance of the organisation. In the ultimate instance success and failure of an
organisation can be explained in terms of the organisation’s ability to manage both positive
and negative constraints – the risk interdependency field. This is the idea embedded in the
discussions that follow through to chapter 6 of this thesis, the idea of organisational
actorhood.
Thomas (2007), Whitley (2008) and Macfarlane (2012) think the professoriate ought to be
more involved in strategy formulation at both the national and institutional levels. This is the
quintessence of actorhood: participation, inclusion and social responsibility. Actorhood in
QFD is about doing these with the goal of improving the organisation’s customer satisfaction
performance. The ideas of participation (and more so presencing), inclusion and social
11
responsibility (feeling beholden to fulfil the Voice of the Customer and the mission of the
organisation) pervades all future arguments on customer satisfaction performance and the
value of QFD. Crotty (1998:3) conceptualises a theoretical perspective as the philosophical
position or stance that informs the methodology and thus provides a context for the research
process. Broido and Manning (2002:434) and Meloy (2008:141) reinforce Crotty by arguing
that the method cannot be disengaged from theory and other accoutrements of pre-
understanding which in their core influence the hermeneutics of interpretation and
representation of what is studied. Figure 2.1 below captures these positions and will create a
basis for future understanding of how QFD can work as a method for programme quality
assurance (PQA). Most models for quality assurance are input-based and focus on the
presence of sets of inputs without interest on assessing and evaluating how the inputs are
actually transformed to meet the needs and wants of the various customers. QFD
conceptualises quality in terms of those inputs, their interfaces, as well as in terms of how
they get transformed and delivered. This characteristic of QFD should give it an edge over
the other quality models. But how well it will do it depends on the perceptions and
competences of those adopting and working the QFD model.
The successes of the Theory of Constraint and QFD within public, private, manufacturing
and service sectors including education are widely ventilated in literature (Nutt, 2002; Vora,
2002 and Bernasconi, 2011). My discussion of figure 2.1 below should help to show the
incremental adoption of the Theory of Constraint and elements of QFD in education. With
this we should be able to assess and evaluate how HEIs ought to manage constraints to
quality and the QFD model. The Theory of Constraint and QFD cover similar areas of
management: operations-production management; strategy and tactics; distribution and
supply chain management; finance and measurement management; sales and marketing
management; project management and people management (Srinivasan et al., 2005). An
overemphasis of one aspect or sticking with archaic assessment tools can become a hindrance
to quality improvement within higher education institutions. Figure 2.1 show the evolution
of the focus of quality assurance models. In its infancy (per Figure 2.1) the control for quality
was centred on the process of doing the job and this was the prerogative of the craftsman.
12
Organisation-
wide quality
control
TQM-the culture
of continuous
improvement
Total Quality
Management
Quality
Assurance with
a quality
Department
Statistical
Quality
Control
Inspection
based
Quality
Control
Quality
control by
the
Foreman
Quality as an
integral part of
craftsmanship
Pre-1900 1900-1920 1920-1940 1940-1960 1960-1980 1980-1990 1990-2000 2000-present
13
After the 1890’s the assumption was that quality should be controlled from outside by the
foreman. There were no modifications that could be featured on products and services
without prior approval by the foreman. As from the 1920s quality began to be inspected by
teams and the inspection would be based on some pre-determined criteria. Workers would
receive training and apprenticeship and would be trusted to produce goods and products that
would meet the wants and needs of the customers. Products with minor defects (rejects)
would be sold at give-away prices. The worst ones would be reworked or salvaged and
destroyed. As industrial production increased and science and technology advanced bringing
new and innovative ways on the market there arose mathematical and statistical ways of
controlling for quality. By the 1960s quality departments began to emerge in companies with
the roles that included studying competition and their products’ sales points with the idea of
building a body of knowledge that would be used to improve and innovate on own supply
and production chains. Up to this point quality assurance had a front-end character.
It was with TQM that quality began to be pushed upstream and focus on both the quality of
inputs as of the throughput processes. A factor that continued to be overlooked was the
philosophy of quality. The value of the non-technical, the human dispositions and mental
modelling were apparently neglected. QFD began to gain prominence with its advocacy for
Voice of the Customer. But still many did not understand what Voice of the Customer was
supposed to mean and today it still remains for most organisations an end in itself to gather
Voice of the Customer.
In thick QFD, Voice of the Customer is but a means, a point of departure and a premise for
escalating the expectations of the various customers into the organisation’s management and
production strategies. In more innovative QFD cultures Voice of the Customer is a means to
understanding the market, not an end in itself or a show of blind compliance. Voice of the
Customer should be translated into Six Sigma roadmaps and escalated to Product Planning
Matrix. Doing this aligns the creation of products and services to the needs of those who will
use the products and services. In deed the strategic value of Six Sigma within QFD is to
enhance the market-orientation of products and services. Having discussed the elements of
the Theory of Constraint the next section looks at the conceptual framework that facilitates
14
the organisation of the assumptions and ideas so that the research objectives are
accomplished.
Figure 2.2: Modified Strategy Focus Wheel (Roberts and MacLennan, 2006:11)
A mentality for continuous improvement leads to mapping an organisational atlas for goal-
based change which feed the organisation’s desire to be out front (the best or leader in the
business). A desire for continuous improvement creates the impetus for a new mode of
strategic planning aimed at market excellence. But market excellence is a function of the
organisation’s excellence in leadership. Consequently market, leadership and production
excellence must be cast within the organisation’s actorhood. This implies framing these
within the organisation’s strategy and every strategic planning activity. The change-project
management infrastructure should roll-out the organisation’s focus on its strategic intent,
show which strategies have priority and how the projects that bring about the desired changes
should be driven through.
17
Making the desired improvement-focused strategies work while focusing on driving the
desired change creates an infrastructure of resources for the near and far future. However the
futuristic nature of part of this effort calls for a robust approach to strategically manage
strategic level, management level, operational level and unforeseeable risks (contraints).
Strategically managing the organisation’s risk interdependency field should help the
organisation set favourable conditions for future improvements and current performativity.
This makes visible the positive and negative constraints within the organisation’s landscape
allowing thus a more richly informed atlas of organisational change. I discuss each strand of
the Strategic Focus Wheel to much greater profundity below.
Continuous improvement sits at the heart of successful innovation and improvement (Al-
Kassem et al., 2013) as it can build a firm foundation on which to edify an innovative
organisation (McAdam et al. 1998). One of the problematic issues with continuous
improvement is that most management think they can demand continuous improvement from
their workforce and that they can tie it in with employee employment contracts. True
continuous improvements must focus on the improvement of each worker, of all processes,
structures and their interfaces. Ignoring workforces’ personal mastery skills, their
understanding of their individual capacity to produce results and mastering the principles
undergirding the way one produces results is a fatal mistake in leadership seeking innovative
change. Management must focus on enhancing a desire for presencing, and a feeling for
effortless and joyousness in every assignment undertaken (Senge et al., 2005; 2010b).
Each principle is effectively the vehicle, the rule and resource that deploys quality within the
strand it belong. For instance strategic planning is deployed throughout the organisation by a
workforce that is pushed by their intrinsic penchant desire to see their organisation emerging
as a reckoned quality performer and an innovative organisation. The basic unit of structure
and functionality of organisational / university actorhood is the staff. It cannot be denied that
staff and academics in well reckoned HEIs equally gain in terms of social status when their
organisation takes on the public limelight. Cognisant of this spill-over gain, staff and
academics can purse strategies that will balance innovation, improvement and quality. This is
the essence of the Strategic Balance Theory (SBT). They too will seek to create
organisational elements that will distinguish it and them from others. This is the essence of
the Optimal Distinctiveness Theory (ODT) (Galvin, 2009). Strategy balancing and optimal
distinctiveness tendencies are not at variance with QFD. I hope we appreciate how staff can
be incentivised to improve the lot of themselves and their institutions by the SBT and ODT.
In figure 2.1 above I said that change-project management help the organisation focus on its
strategic intent or what it needs and wants achieved. In table 2.1 I show that this effort is
deployed throughout the organisation through the management / excellence principles of:
integration of effort; being disciplined; creating customer value; being time-based and
creating strategic capabilities. From figure 2.1 we observe that making strategies work is a
special endeavour directed at getting the proposed change-projects work. This process is
resourced and catalysed by a set of rules or principles that guarantee structure-structure,
20
structure-function and function-function alignment; embracement of change; establishment
of a learning culture; relating the micro to the macro; measuring and reporting; supporting
distributed leadership and being up front (honest, trustworthy, moral and ethical). Strategic
risk management is about creating the resources that will support the organisation’s
prosperity. This is about watering, fertilising and ensuring the desirable characteristics grow
and spread whilst weeds are removed. Resourcing for the medium term is one vehicle that
deploys this quality.
Table 2.1 expands on Figure 2.1 by embracing the principles or rules by which each of the
Strategy Focus Wheel’s strands and each of the segments by Bevington and Samson (2012)
is deployed across the organisation. Notice how the brief descriptions relate vertically and
horizontally. I explain each level below, however using only strands of the Strategy Focus
Wheel.
Of the 14 management Best Practice Principles the one that best plugs in with the ‘Strategic
Planning’ strand of the Strategy Focus Wheel is ‘‘Having the desire to be out front’’ and
being the best in the park (see Table below). A desire to be out front with every strategic
element of higher education like staff quality, programme relevance, delivery modes and
rigour of content creates a set of strategic capabilities that matter the most in market-oriented
education.
21
Table 2.1: Relation between Strategic focusing and the Best Practice principles as synthesised from literature
22
Strategic planning is about having a penchant desire to achieve something, it is about having
an aspiration and an ambition and the glue to draw into the mainstream the men and women
who traditional bureaucracies would keep in the peripheries of decision-shaping
(Dervitsiotis, 2002:1087). Wei et al. (2015) observed that strategic planning with a desire to
be out front is crucial for ongoing instructional improvement.
23
not likely to mature. Five of the 14 management best practise principles that relate most to
change-project management are:
Ensuring integration of effort
Being disciplined
Creating customer value
Being time-based
Creating strategic capabilities
Ensuring that the organisation has the ability to manage change guarantees the organisation’s
ability to co-adapt with the changing internal and external environments particularly within
the supply chain and in the customer base.
Wider discussions and alignment of business processes could help to displace turf warring
and fragmentations between higher education institutions and their quality assurance
agencies. Senge et al. (2007 / 2012) and Bevington and Samson (2012:177) observed that turf
warring, fragmentation, reactivity and competitive behaviour lead to organisational
dysfunction. A turf war is generally a struggle for power, control, and other social goods such
as recognition among an organisation’s stakeholders. Disconnects among students,
academics, management, quality assurance agencies and society are widely blamed for poor
quality performance of higher education institutions. They are potential causes of turf-
warring. In chapter 3 four Six Sigma roadmaps are discussed and their importance in helping
to close these disconnects in favour of integration and alignment for quality is underscored.
Organisational discipline is one omnipotent tool for establishing alignment and integration.
24
2.3.5.2 Being disciplined
Policies and organisational structures that are blindly adopted from other organisations would
rarely be as effective in the adopting organisation. Interface mapping based on the voices of
those who do the work should help to shape policies that people can relate to, thereby
appreciating how these policies influence their working environment. Staff generally resist
policies that they find oppressive and alien to their culture. Strategic categorisation would
help organisational management to identify the kind of work and activities that do not add
value to the organisation and harm or are indifferent to customer satisfaction. Strategic
categorisation also helps the organisation to understand and monitor the core activities that
are value adding (Bevington and Samson, 2012:179). Oakland et al., (2002:1132) explain the
importance of a disciplined approach to establishing links between organisational
philosophy, mission, values and the core processes at the ground level. Being disciplined
helps in focusing the organisational infrastructure on creating customer value.
To successfully implement the principle of creating customer value it requires that the
organisation should have performed well with the generation and institutionalisation of a
relevant and enabling policy regime (being disciplined). Furthermore, institutionalisation of
an enabling policy regime that aligns structure-structure, structure-function and function-
function relationships (integration of effort) in the organisation creates a strong strategic
capability in the organisation. Enhancing the organisation’s capability of creating customer
value involves interface mapping which is a strategy by which integration and alignment can
be achieved (Bevington and Samson, 2012:180). Only time-based organisations would
25
realise the value of interface mapping, which actually helps in removing any time-wasting
activities of no value to the products and services. In fact, redefining time as a critical
resource enhances organisational efficiency and optimises the creation of customer value.
This is the essence of being time-based.
Literature has many similar examples of organisational responses to the realisation of the
value of time and trimming organisational activities to only those that are critical to
customers. These realisations have fathered principles such as ‘just-in-time’ delivery;
management-by-objective; performance management and many more. Successful
applications of these principles should proffer organisations the benefits of structural agility,
cultural agility, process agility and change agility. In chapter 2 we learnt that these are
reasons why most organisations, then and now, adopt QFD. QFD caters for all these elements
of organisational performativity. Mukherjee and Wong (2011:138) are of the opinion that “an
approach with a delay built into the process has the disadvantage of impairing both efficiency
and institutional agility to respond swiftly to change”. In higher education delay on delivery
of customer value may be caused by time and efforts spend on waiting for management
approvals, lots of paperwork and sustaining bureaucracy. Bernasconi (2011:243) lambasts
26
public administration rules for inhibiting HEIs from responding with agility to external
opportunities. Cruickshank (2003) laments that even after many years of adopting business
quality models, HEIs continue to be sluggish because, among other factors, there is lack of
vision of the institution as a system of interrelated parts that should otherwise be attentive to
incidents through coordinated responses (Gallifa and Batalle, 2010:162). Mukherjee and
Wong (2011:140) tend to suggest that autonomy helps to bring agility to strong world-class
universities. But the truth is that autonomy is not antithetic to sluggishness. Agility should be
felt as a means to creating the competitive responsiveness that characterise success in
competitive markets. Autonomous Faculties in HEIs may still inherit the sluggishness of the
mother University, for example, unless they are consciously set to be agile. QFD would
improve on time usage because of its emphasis on integration, alignment, leadership
distribution and creating strategic capabilities for improved customer values.
Most assumed change, a lot that would benefit educational quality never set off because
organisations continue to indulge in the past and that there is no mechanism of setting on the
change process and catalyse it. In QFD the underlying assumption is that systems are
improved by the men and women at the systems’ coalface. This implies that more open-door
and democratic management in HEIs that value the voices of their professoriate and students,
of their external stakeholders by genuinely incorporating them in their strategies reward the
institution with a repertoire of robust strategic capabilities and the power to make their
strategies work.
27
2.3.6 Making strategies work
Making strategies work is an ongoing process for connecting the high-level strategic plan
(macro) to the day-to-day (micro) activities that are critical to the delivery of the strategy.
These activities include all secondary changes necessitated by changes in the initial premises
of the strategies. Of the 14 management best practice principles that make strategies work are
the following seven:
Gaining alignment
Embracing change
Establishing a learning culture
Relating the micro to the macro
Measuring and reporting
Supporting distributed leadership
Being up front
Deploying the above principles throughout the activities of HEIs helps to create conditions
for a strategically bundling management, the professoriate, students, instruction, curricular
and the mission and vision.
The communicative, collaborative, and participatory nature of QFD creates conditions for
shared visions. In drawing an atlas for change it is best to start by asking staff to document
fully what they do and their assessment of its implications on value creation, value
management and value delivery. This process creates the first instance the staff comes head-
on with focused self / role evaluation. It should then be easier to move with the staff to the
next level of evaluating what needs be changed by removal, modification or addition. This is
the first instance of interface mapping, a process that feeds into the principles of time-based,
alignment, creation of customer value, integration and of linking the micro to the macro.
Used well, QFD should help with problem identification and design of appropriate change
and solutions. Throughout the text the thesis discuss QFD as a mental tool that facilitate the
understanding of data generated from within as from outside the organisation. It is indeed a
potent tool for organisational learning and the subsequent mapping of an atlas for change.
29
2.3.6.3 Establishing a learning culture
Interface mapping, strategic categorisation and the use of genomic frameworks set the
foundation for a focused mechanism for promoting and targeting learning. Establishing a
learning culture under the auspices of QFD creates what Pearce and Robinson (2009:364)
define as an:
‘‘... organisation structured around the idea that it should be set up to enable
learning, to share knowledge, to seek knowledge, and to create opportunities to
create new knowledge’’
Ficalora and Cohen (2009:4) say ‘‘QFD is a method that flags gaps in knowledge, capability
and understanding as the design team works through the various QFD elements’’. Better
aligned and lean team-based organisations learn better and faster than expansive
organisations (Bevington and Samson, 2012:183). By pinpointing the organisation’s
shortcomings and proffering strategies for their correction, QFD becomes a more
comprehensive knowledge and quality management tool. Knowledge takes numerous forms
which include operating know-how, knowledge of customer networks or relationships with
customer networks. It can also mean technical knowledge upon which processes are or will
be based and the technical knowledge upon which products and services are or will be based.
It can be about relationships with key people or an individual that can get things done more
effectively, efficiently and speedily (Pearce and Robinson, 2000). All these forms of
knowledge help in creating improved educational institutions and programmes.
30
At Section 3.5.3 the thesis discusses four strategies of aligning organisational culture to
business strategy. Benchmarking and networking that are designed to leverage and enhance
knowledge should flourish the more the learning culture calcifies (Harvey and Williams,
2010b:81). Temponi (2005), Trivellas and Dargenidou (2009) are also of the opinion that
business schools should be shining examples of learning organisations where continuous
improvement, customer satisfaction, and service quality are unquestionably paramount.
Yorke (2000) had earlier highlighted that effective learning organisations depend on the
recognition of good ideas and expertise for what they are and not for where they are from or
who in the organisation backs them. Yorke further argues that open support for capacity
building in ambiences of trust, honesty and respect enhance effective organisational learning.
From an organisational management perspective in higher education, one would quickly
assume that business schools should take the lead in exemplifying learning institutions. They
would be characterised by continuous improvement, service quality and customer satisfaction
(Trivellas and Dargenidou, 2009; Temponi, 2005). In chapter 3 I discuss how organisational
and strategic structures that marginalise one or so stakeholders limit the number and value of
options at their disposal.
Literature is awash with tools that help in linking the micro to the macro. These include
employee empowerment; team-based working; Voice of Customer; Management by
Objectives; the Balanced Score Card; customer surveys and benchmarking for management.
These tools, at least on a theoretical level, raise the workforce’s awareness of the strategic
goals. A more efficacious and precise mechanism of linking the micro to the macro, the short
term to the long term and the ‘smaller activity’ to the global, is through interface mapping,
31
strategic categorisation and the use of the ‘genomic’ framework. The implication of interface
mapping is that all routine and seasonal activities are documented and areas of ritualism,
duplications and noise are pinpointed (Anderson, 2006; Bevington and Samson, 2012:182).
This is an important element of QFD implementation particularly in HEIs that are laden with
valueless workloads.
BCIs can be set at the person, team, sector or organisational level and may cover particular
set of knowledge, attitudes, skills, understanding and beliefs. Bevington and Samson
(2012:187) say that ‘‘... when all the BCIs are adhered to, the strategic outcomes will be
delivered’’. This point is reiterated by Dahlgaard and Dahlgaard (2002:1071) who exhort the
need for linking KPIs to training provisions. Traditional quality assurance schemes have the
weakness of collecting hard-and-dry data without tracing backward as to causes or drivers of
the data and weaknesses. In QFD contexts, data is collected with the idea of building
effective action. This means BCIs are subordinated to the organisation’s key performance
goals and indicators, thus reinforcing cross-sectional integration and both vertical and
horizontal alignment. Measuring and reporting is a powerful QFD tool for building a shared
organisation-wide understanding of where the organisation stands, its fitness for the next
challenges and its risk envelope. The shared understanding is critical for creating incentives
for distributed leadership.
32
2.3.6.6 Supporting distributed leadership
Further to empowering the workforce, the practice of distributed leadership engages staff in
analysing the business with an eye to enact necessary changes and in relating their daily
activities to the medium- and long-term goals of the organisation. Distributing leadership
goes beyond assigning job descriptions. It is about understanding why one has to do certain
things and accepting responsibility for improving one’s performance, such as making
operational decisions pertaining to one’s roles in the team and the organisation. Genuine
leadership distribution involves being up front.
In QFD contexts distributed leadership should aim at developing a robust infrastructure for
leadership substitution. By leadership substitution is meant a context in which subordinate
professional orientation is so honed that they need no coercion to do their roles. It further
means an exceptional amount of competence that supervision become minimal. Jones et al.
(2012:67) observed that leadership that befits ‘‘... the higher education sector …’’ is one that
‘‘… requires a less hierarchical approach that takes account of its specialised and
professional context’’. Pierce and Newstrom (2000:254) view employees with a professional
orientation as being able to:
‘‘... cultivate horizontal rather than vertical relationships, give greater credence to peer
review processes, however informal, than to hierarchical evaluations, and tend to develop
important references external to the employing organisation.’’
Corruption and favouritism constitutes a negative risk that exists if management control over
processes is not sufficiently vigilant (Fill and Jameison, 2006:38). In HEIs, the most
ventilated incidents of unethical and immoral behaviours include misuse of institutional
property, sexual exploitation, research plagiarism, bribing of teachers, cheating by students,
and irregularities in the admission and examination processes (UNESCO, 2007:15;
Heyneman, 2011:8). Corruption tumbles the reputation of HEIs known for perpetrating it and
reduces the competitiveness of their graduates on the labour market. The Transparency
International (TI) calculates a Corruption Perception Index (CPI) that reflects on the levels of
corruption, opacity and lack of meritocracy in a society (Jabnoun, 2009:421). Cognisant of
the negative impact corruption in HEIs will have on future employability prospects, youth
organisations in South East Europe founded the Anti-corruption Student Network in 2003.
Their objective was to coordinate the exposure of immoral and unethical practices in higher
education. This effort is being replicated in many HEIs. The Higher Education Corruption
Monitor proffers information from news articles, bibliographies, and links to other agencies.
There are other strategies at play to curb corruption and lower the perception that a HEI
could be corrupt. Heynenman (2011:9) lists some of the strategies:
annual reports to the public on year-by-year changes in corruption incidents
code of conduct for administrators and staff
code of conduct for faculty
code of conduct for students,
34
exposure of perpetrators
institutional courts to hear cases of misconduct
staff surveys on corruption
statements of honesty on public websites
student surveys on corruption.
Academic corruption becomes all the more difficult to deal with because immoral, unethical
activities carry cultural, legal and social connotations. And the chances of corrupt behaviour
increase in high-power distance communities where checks and balances are difficult to
implement effectively (Kimbro, 2002; Sanyal and Subarna, 2004; Dirienzo et al., 2007).
The second rule-resource relationship is when these same prospector organisations valuing
knowledge (strategic capabilities) for its own sake (rule) create power to define problems
(resource). They view every team member as a knowledge-node, a leader and a distributer of
knowledge. Organisations that live by the rule of sharing responsibility and integration for
the entire project, proffer themselves with the resource of time and attention.
Non-innovative organisations on the other hand have sets of desanguiniating rules and
resources that haemorrhage workers’ creativity and innovative ideas. Whichever model gets
adopted in such organisations, there will be minimal success. Non-innovative organisations
36
are characterised by the pervasive presence of the following rule-resource relationships. A
non-innovative organisation is overwhelmed by its own focus on results thus generates the
resource of control at the expense of expertise. Power, in them, is more important and is only
with relation to getting a fixed list of results. This is characteristic of many a quality
assurance agency and the result has been ‘comply, comply, comply’ when they come to
evaluation of institutions. This maybe is one of the reasons institutions have remained stuck
with externally defined metrics and standards that are at variance with the needs and wants of
the end users of their products and services. In non-innovative organisations the rule of
eliminating problems quickly generates the resource of limiting discussions and thus limiting
others’ actions. This is in stun contrast to QFD and the desirable way of running educational
projects in competitive environments. In trying to keep high results or to exonerate self from
poor performance by saying ‘‘we don’t know how it went wrong but we followed the rules’’.
Non-innovative organisations discard any suggestions to see or do anything differently.
The rule of separating responsibility generates the resource of control over one’s own
domain. This rule eliminates the possibility of cross pollination of ideas and has no value in
teamwork quality. Where people work in numbers it is about the physical gains that result in
net ability to lift the load or to reduce completion time not about adding value and
satisfaction.
QFD is characterised by the first three sets of rule-resource relationships. The second set
tends to characterise the very traditional HEIs and quality assurance agencies. The reason for
adopting a hybrid model of Strategy Focus Wheel, the 14 Best Practice Principles and the
Rule-Resource model is that neither, standing alone, would give a fuller explanation of why
humans would behave the way they do in the face of a change-project model like QFD. And
by using all three perspectives the thesis maintains the complexity of institutions and the
complexity of their management. The 14 best practice principles expand the concept of
‘Strategy Focus’ and are the vehicle that deliver excellence in service and product delivery.
The Rule-Resource model explains what distinguishes winners from losers, even with the
same model. The elements of the proposed model are thus interwoven, as shown below. As
each plugs into the other it is difficult to present a model of a conceptual framework that
37
exhibits mobile relationships pictorially. As chapter 2 continue to unfold, the key point is
imagining how the management excellence principles could intervene to deploy quality
throughout the organisation’s structures, processes and functions. But we need to discuss the
perspectives of QFD in order to build a corpus of concepts for analysis of why the model
would be an urgent adoption.
2.4.1 The quality perspective is defined in terms of the qualities, attributes or features
making the expectations of the customer. Literature generally highlights some ten
determinants of service quality which (Franceschini, 2002:48 / 147-’8) use to link perceived
quality and expected quality of a service. These determinants are: access, communication,
competence, courtesy, credibility, reliability, responsiveness, security, tangibles and
customer knowledge. The quality perceived by a customer is a function of the complex
relationship among the quality factors themselves, word-of-mouth communications, personal
needs and past experience. Figure 2.3 below shows how the ten determinants of service
quality (Franceschini, 2002) relate to perceived service and expected service in building a
perception of service quality. Whilst we can list the determinants of quality discretely our
mental models of quality don’t isolate them but pass a complex picture of what they
represent as quality to us. However each of the determinants percolates into our mental
models individually or maybe in permutation with other determinants based on our
exigencies, past experiences and / or internal and external communications. Three levels of
satisfaction would arise independently of one another but expressive of the balance between
expected and perceived quality.
38
Figure 2.3: The relationship between ten determinants of quality, perceived and expected
quality of a service (Fransceschini, 2002).
Perceived quality has a mobility characteristic just as each of the ten factors is mobile and is
historically conditioned in space and in time. That means ‘relational, organisational and
temporal variables of a service’ must be defined and evaluated. This point is of fundamental
value particularly in curriculum design and the ensuing teaching-learning interactions
Franceschini (2002:139).
Babakus and Mangold (1992), Chua (2004), Tan and Kek (2004), Mahapatra and Khan
(2007), Narang (2012) confirm the use of SERVQUAL in measuring quality in HEIs. The
SERVPERF has also been used in education with some claimed success. However, these
models have their shortcomings which propagate into many other models derived from them.
Narang (2012:361) proposes the use of a model, EduQUAL which is expressed
mathematically as:
Other authorities have looked at quality variably. Harvey and Knight (1996) talk of quality as
perfection / excellence, a construct that relates to the principle of being right the first time.
However, there are difficulties in defining what is excellent. Attaining excellence in every
facet of a product or service may not be possible when looking at the situations regarding
context and resources in which most organisations operate.
40
Quality as consistency to some specifications may be proper in manufacturing, but a rather
naive pursuit in the education delivery system. With reference to the services sector Oakland
et al. (2002:1126) refer to consistency of material, skills, facilities etc. Consistency may be
better looked at as the alignment among inputs, processes and outcomes (Stufflebeam and
Shinkfield, 2007) . This is shown in Figure 2.4 below. The basis of the model is that the
context defines quality and the inputs and throughputs conform to the conception of quality
held. In this way quality offered approximates quite closely to quality planned. Elements of
the model still persist in novel models for quality assurance, including QFD. This context,
input, process, product model assumes that both the intra- and extra- institutional
environments (context) partake in determining inputs, throughputs and standards for
measuring outcomes as satisfactory or not. Since contexts differ in space and in time, it
means that the nature of inputs, throughputs and standards would differ too in space and in
time. However in QFD the basic inputs are the voices of the diverse customers. The main
goal for outputs is their fitness for purpose as measured by Customer Satisfaction
Performance metrics. Figure 2.4 below summarises these points.
41
CONTEXT
Government
. Other
Industry Society
Quality
Figure 2.4: A QFD-based model for insuring deployment of quality throughout inputs, processes their interfaces and outcomes
42
A model must recognise the context in which an assessment or evaluation is being sought
and done. The aim in good education is not to ‘drill students’, but to work with them within
their academic, social and spiritual diversity. Not even is the aim of education to produce
‘cloned’, like-minded graduates. With education we may look at consistency with ‘Voice of
Customer’, with the university mission and the university’s vision; with the skill sets
necessary for a prosperous life; and which are consistent with democratic, peace-loving
societies. Unlike in goods production where consistence is about exactly similar goods, in
services consistence is in respect of Voice of the Customer, the mission and vision of the
organisation.
For Macfarlane and Lomas (1999:70) quality as value for money means that graduates
should be able to return on the investment through development and innovation. To Lemaitre
(2009:3, 2012) it means technical efficiency, while for Hertzman and Ackerman (2010:211),
Law (2010a:66), and CHE (2011:13) it means accountability, effectiveness and efficiency.
These positions co-relate with the Six Sigma notions of designing organisational structures
and processes that reduce failure modes. Ewell (2007) looks at quality as value for money in
terms of rigour of study programmes and employability of graduates. However
Meirovich and Romar (2006:328) warned that ‘value for money’ should be cushioned from
students’ pressures for passable examination and inflated grading. Lamentably, some quality
assurance instruments place an emphasis on retention and graduation rates, for instance the
ZimCHE. These measures can be easily manipulated and are unrelated to quality as
transformative and as fitness for purpose.
Quality as transformation has been discussed by Harvey and Green (1993), Harvey and
Knight (1996, ch. 1 & 2), Richardson (2000), Tam (2001), Harvey (2002), Law and Meyer
(2010), who concur that this refers to value addition on student knowledge base, attitudes,
skill set, understanding, behaviour and belief (KASUBB) system. The implication here is that
quality should be defined fundamentally as what happens in the teaching-learning
interactions. It therefore collates well with educational evaluation models that focus on
inputs, processes and outputs. I consider ‘quality as transformation’ a super-model that builds
on yet other models: Engagement model of Program Quality; University of Learning model;
43
model for a Responsive University. In proposing the engagement theory, Haworth and
Conrad (1997) argue that an institution of learning would do more and better at improving
future value of their graduates if they invest in: programmes dependent on ‘Voice of
Customer’; diverse and engaged stakeholder groups; participatory cultures; interactive
instruction, and adequate resources. The university of learning model (Bowden and Marton,
1998) defines functionality of the university in terms of the same characteristics that Hoegl
and Gemuenden (2001, 435) would use to define ‘teamwork quality’: communication,
balance of member contributions, mutual support, effort and cohesion. The responsive
university model (Tierney, 1998) presumes that the university should be responsive to
‘voices of customer’ and be service-oriented, and emphasises the same facets of ‘teamwork
quality’. The three models reinforce the concept of university actorhood referred to earlier.
According to Campbell and Rozsnyai (2002:132), ‘quality as fitness for purpose’ presumes
that the education processes are subordinated to a datum of university vision, mission,
customer and other evidenced requirements, and that the graduate will give evidence of the
fulfilment of purposes of higher education in its broader sense as in the narrowcast sense
implied by the programme’s curriculum. The transformation done by the teaching-learning
process must improve students’ knowledge, attitudes, skillset, understanding, belief system
and behaviour (KASUBB) so that they: become research professionals, employers,
competent employees and creators of goods and services that improve society’s wellbeing.
An argument is made here that the concepts of quality as ‘consistency’, ‘value for money’,
and ‘excellence’, hard as they are to measure, form the foundation for the construct of
‘quality as transformation’, which in itself leads to ‘quality as fitness for purpose’. Looking
at the concept of quality in this integrative manner helps to define and orient the argument for
outcome-based curricular and objective-based education.
2.4.2 The function perspective refers to everything that has an influence on the ultimate
quality of products and services provided by the institution. This includes students,
stakeholders, staff, management, processes, resources, physical infrastructure, teaching and
learning material, the appropriateness of the interfaces between the institution’s structure-
structure, structure-function and function-function relationships, etc. It also refers to the
44
principles, values and tools that the organisation deploys in its multiple activities. It further
refers to the management of networks and external strategic architectures, among other
things. This perspective stretches the traditional collegiality concept of the ‘teamwork
quality’ concept, which is the hallmark of QFD. Most quality assurance schemes tend to put
lots of focus on issues that fall within the function perspective of QFD such as scores of entry
students, academics’ credentials, infrastructures, etc. This action is incomplete in the sense
that real quality arises from the correct deployment of the functions and their strategy-
operational alignment.
2.4.3 The deployment perspective defines how the flow of development efforts will be
developed, deployed, diffused, and managed to ascertain that they are aligned or market-
oriented to the needs, requirements and expectations of students, society and industry. It
refers to how the ‘needs analysis’ is translated into the skills and processes of the institution
so that the programme achieves market orientation (Ficalora and Cohen, 2009:176). This link
finishes one of the most important matrix in QFD – the whats-how (objective-means) matrix.
The overarching statement on deployment is how tasks, resources, efforts, time, recognition,
power and responsibilities are allocated or deployed among structures, processes and persons
within the institution. Relationship matrices, open and sincere communication become of
such great importance that alignments are achieved and that performers are clear as to why
particular decisions are enacted. In QFD quality is deployed through management of the 14
best practice principles across every function and interface in the organisation. Building on
the arguments under each perspective, the thesis assumes QFD as the translation of Voice of
the Customer into quality objectives that are then deployed to appropriate structures and
functions of the organisation.
QFD is a problem-solving tool. Franceschini (2002:13) and Ficalora and Cohen (2009:xviii)
are of the opinion that QFD proffers a formal linkage between each stakeholder requirement
or need and the organisation’s responses to each of these needs. This linkage is achieved and
communicated through processing Voice of Customer and constructing the relevant matrices.
Elevating ‘Voice of Customer’ to Six Sigma roadmaps as shown in Figure 3.10 helps
organisations to solve issues relating to demand-resources-competitive performance. For
instance:
Design for Six Sigma (DFSS) focalises on overall organisational response to customer
requirements. The idea being to adopt an organisational structure, behaviour and
mentality that focus on market excellence through effectiveness and efficiency.
Six Sigma Process Design (SSPD) focalises on (re)designing person-to-person, person-
to-system and system-to-system processes with the aim of making the organisation lean,
effective, efficient and agile by removing time, people, and resource wastages.
Technology for Six Sigma (TFSS) focalises on the adoption of appropriate technologies
for more effective customer-driven performance, whilst
Marketing for Six Sigma (MFSS) focalises on up front communications with the market
with the aim of customers and stakeholders sharing the vision and mission of the
organisation.
Matrices are used all over QFD to help the stages link goals to means. This is where and
when those who are unsettled by matrices claim that QFD does not work. Ficalora and Cohen
(2009:xviii) affirm that QFD offers a roadmap for assessing and evaluating the best strategic
options for meeting customer needs with resources at hand, irrespective of the technology
underlying the products and services.
47
QFD has been described as the visible memory of the organisation and as a method that
provides a convenient repository of the organisation’s information right from Voice of
Customer to management-related documents. This information can always be retrieved,
reprocessed and re-membered with other data for new decision-making processes. Matrices
however constitute a high-level summary of key product planning data in the repository
(Ficalora and Cohen, 2009:9). By determining the data and its deployment in Product
Planning Matrix and strategic planning and detailing all future amendments thereto QFD
writes a concise history of the organisation.
In conjunction with this function, QFD proffers a systematic method of comparing and
setting project priorities at different levels of the organisation with respect to the products
and services development processes. One of the greatest problems in QFD implementation is
when management fails to realise the supremacy of Voice of Customer over their intuitions.
Ficalora and Cohen (2009:4) are of the opinion that ‘‘QFD enables management to evaluate
whether the product plans are worth the investment’’ as well as take appropriate budgetary
decisions, including allocations of other resources to activities. Every too often literature
blames HEIs for lack of rigour and relevance in their programmes. An equal amount of
research blames the Quality Assurance agencies for failure. HEIs respond by instituting
changes that sooner or later become an attraction to yet another barrage of criticism. Where is
the problem?
Ficalora and Cohen (2009:9) believe that in the hands of the organisation, the team or the
individual, QFD informs the development or deployment of means based on objectives at
every altitude of the organisation. QFD assumes that task deployment as well as skill /
response deployment should be in pursuance of maximising on each objective, which in itself
should arise from the overarching organisational aim. A problem, particularly with novice
management, is role-induced biases in decision making rather than objective, fact-dependent
decisions.
QFD is basically a strategy design tool that helps greatly in aligning goals, means, and
resources. QFD is used as an organisational alignment tool. Ficalora and Cohen (2009:4) say
48
of QFD that it ‘‘keeps track of how key product and process design decisions relate to
customer needs’’. As QFD is taken through its stages, it flags gaps in knowledge, capability
and understanding. By taking track and highlighting constraints, QFD helps the
accomplishment of intra-project alignment as well as the much-needed alignment between
management desires and project rendition. It also influences decisions on the amount of
support required. Ficalora and Cohen (2009:4) argue that ‘‘QFD is fundamentally a quality
planning and management process to drive the best possible products and services
solutions’’.
Akao (1990), and Franceschini (2002:22) consider QFD to be a ‘mind enhancer’, ‘... an
evident and powerful tool for prioritising and assigning’ or deploying resources and
responsibilities throughout the organisation. QFD is widely used to link customer
requirements to skills development programmes, to work breakdown structures (WBS) and
organisation breakdown structures (OBS). This implies that quality assurance functions need
to be in place well before the products and services creation processes begin. This shift of
quality assurance functions, from output stages back to the pre-production stages, is one huge
contribution of QFD practitioners of the 1960s. Salmi (2011b:6), writing on universities
seeking world-class status, reiterates the value in having leadership, the board and a robust
strategic framework well before programmes and projects begin.
Franceschini (2002:39) says QFD is being used to prevent interface drawbacks and improve
the organisation’s horizontal integration and leveraging. Akao (1990), Oakland (2000:39),
Almannai et al., (2008:4) and Ficalora and Cohen, (2009:262;) view QFD as a methodology,
whilst Zheng and Pulli (2007:370) refer to QFD as a philosophy. Sallis (2002:3) however
considers QFD to be both a philosophy and a methodology.
‘‘QFD is what QFD practitioners do,’’ say Ficalora and Cohen (2009:8) and as a versatile
tool / methodology the profundity of application of each of the QFD techniques and tools
depends on the benefits the organisation or team needs or is willing to work for. Thus QFD
should remain open to be implemented at a level of detail commensurate to the task at hand.
With this said, this thesis assumes three levels of adoption of QFD:
49
Dogmatic adoption that is blindly adopting the model without contextualising it, thus
forcing the QFD model onto the organisation. This approach is normally coercive, taking
the logic of instrumentality and consequences. A dogmatic approach will need to be
supported by a barrage of regulative antecedents: policies; rules; regulations; deadlines;
and swamps of sanctions. Here the engine works, but the vehicle doesn’t go anywhere.
Selective adoption may be either dogmatic or incremental, but the organisation chooses
tools from the QFD model and then adsorbs them into the organisation’s modus operandi.
In selective dogmatic adoption, the organisation adopts selected parts of the model in a
dogmatic, non-contextualised fashion. In selective creative adoption, the organisation
adopts selected components or tools of the model in a creative way to fit in with the
contextual elements of the organisation.
In a summative manner, organisations are using QFD along its three vectors. Procedurally,
QFD uses a series of quality tables and matrices. Behaviourally, QFD assumes that everyone
contributes to quality by exhibiting cooperative, collaborative and communicative behaviour
at all times. Philosophically, QFD works optimally when underpinned by results-oriented
management like ‘management by objectives’ and ‘management by processes’. The
philosophical dimension focalises on both what needs to be done and how it is to be done
(Conti, 1989, Lomas and Nicholls, 2005).
50
2.7 Customer: The focus of a QFD strategy
In Six Sigma and QFD lexicon, the term ‘customer’ encompasses a client, buyer, vendor,
purchaser, acquirer, receiver or giver as in organisational supply chains. It means therefore
that in an organisation everyone is a customer to someone. Lomas (2007) discusses whether
it would be appropriate to regard the student as customer to her host institution. In a
competitive market it benefits the organisation to focus on active customers, not-customers
and non-customers. This helps in structuring the organisation’s marketing and quality
strategy. Most research on institutions adopting ‘new public management’ (NPM)
approaches conceive the student as the primary customer. This research adopts a similar
position. It looks more reasonable under the lens of QFD to regard the student as the primary
customer because he is the main recipient of the educational products and services and also
the one who pays the consideration for the education provided. In her study, Ermer (1995)
considered students, academic staff and industry as customers and she analysed their
requirements separately. Samford University defines its customer as the student. Harvard
University seems to take a similar position defining customer as anyone to whom it provides
information or service, which is exactly what the student gets (Hubbard, 1994). While other
HEIs are not too explicit on taking their students as customers, the Oregon State University
says “our students are our purpose for existence” (Coate, 1990). Helms and Key (1994)
reported that students at the Wright University overwhelmingly and jealously didn’t want to
share the title of customer with any of the University's stakeholders. Similarly, Sharabi
(2010:323) agrees that the ‘defining’ of students as customers in HEIs has become
commonplace. The University of Pretoria refers to students as clients and many people use
the terms client and customer interchangeably. This study takes a customer as any party
receiving a benefit for a consideration, and this qualifies every stakeholder to HEIs as a
customer. It also accepts the discriminate ranking of customers.
There were two prominent drivers for QFD in the 1960s. Firstly, quality charts and models
were used in the manufacturing, but on the finished product rather than at every point prior to
churning out the final product (Ficalora and Cohen, 2009:20). It was like worrying about
closing the stable door when the horse has long bolted out. On the other end, measures of
quality in education have always been on the input side: the curriculum; the magnificence of
physical infrastructures; qualifications of the teachers; finances of the institution, etc. What
was wanted was a strategy of quality inputting, quality management and quality delivery and
not of inspecting for presence of quality in services already on the market. Secondly, people
had some idea of the value of design quality, but were unclear about how this could be
incorporated into the production process (Akao, 1997).
What actually happened was that production and quality assurance were disjointed processes.
Quality assurance would wait until production was through, then would come in on the
finished products and services and if anything was wrong research to establish the causes of
failure modes would then be instituted. This approach tends to violate the excellence
principles of ‘being time based’, of ‘integration’, of ‘alignment’ and linking the micro to the
macro. With regard to these drivers, Dr Akao had wondered why production management
52
and doers could not note the critical points on the quality control process chart as
predetermined control or check points for production activity before production start-up
(Akao, 1997).
At the time, quality control charts had the deficiency of confirming quality only after product
or service generation had began (Ficalora and Cohen, 2009:20). In essence, QFD arose to
solve the problems created by quality control mentality and procedures that were skewed on
the output at the neglect of the input and the throughput. This is not uncommon even today.
Many quality assurance strategies are input based. The ZimCHE in its quest for quality
within HEIs proclaims that every university lecturer must be a PhD holder by 2015.
Throughout the years, Japanese quality protagonists continued to exchange details on tools,
templates and techniques of quality. In 1966, Kiyotaka Oshiumi, working at Bridgestone
Tyre Corporation, constructed a Process Assurance items Table (PAiT) (Oshium, 1966). The
PAiT highlighted the transition from true qualities to substitute quality characteristics to the
process factors. This is a crucial step in the QFD’s House of Quality (Ficalora and Cohen,
2009:113). A mentality of bringing in ‘Voice of Customer’ into every step of products and
services creation right up to creation was eventuating. Benchmarking on the PAiT, Akao
worked out a ‘design viewpoints’ field which he blended into the PAiT and made an effort to
have the enriched version of the PAiT utilised in production manufacturing. The new version
53
of PAiT was first described as quality deployment (QD) in 1972. The inception of the PAiT
marked the marketing of a method to deploy, right at the product design stage, the critical
quality assurance (QA) points required to ensure and guarantee the design quality throughout
the production process (Akao, 1997; Ficalora and Cohen, 2009:20-21). At this point, tools
such as factor maps, tables and matrices drifted into the centre of QFD practices. Throughout
1973, Shigeru Mizuno and Yasushi Furukaya working at Kobe Shipyards – Mitsubishi
Heavy Industries mentored improvement of the extended or enriched PAiT so that it could
systematically reflect true customer quality needs in terms of functions, exhibiting the
relationship, the functions and the substitute quality characteristics (SQC) (Mizuno and
Akao, 1994; Singh et al., 2008:164). The integration of the ideas and models so far became
known as ‘quality deployment’ and defined as:
“…a methodology that converts user demands into substitute quality characteristics,
determines the design quality of the finished good, and systematically deploys this
quality into component quality, individual part quality and process elements and their
relationships’’ (Akao, 1990).
Katsuyoshi Ishihara made an invaluable input to QFD by applying concepts from Value
Engineering. Value Engineering had a model for defining functions of a product, and
business function deployment was articulated to the then narrowly defined QFD.
54
2.11 Application of QFD in education
Okamoto and Rioboo (2002:1) observed that education systems have not exempted
themselves from:
“... implementing improvement strategies that can lead them to better quality, reduced
cost, reduced delivery time, as well as better design of services and products, supply
chain management, strategic planning and project management.”
In his editorial to the Journal Quality Assurance in Education, Dalrymple (2010) exhorted the
protracted innovative application of approaches and methods from other discipline areas to
the quest for measurement and improvement of quality in the education sphere. Among these
migrant approaches and methods are QFD, Total Quality Management (TQM), Failure Mode
Evaluation Analysis (FMEA) and others that have become common in educational
programme quality assurance (PQA). Ever since the prowess of QFD was experienced in the
industry, QFD has assumed an expansive content and context, developing into a philosophy
and methodology of doing work.
Krishman and Houshman (1993) found that QFD was an effective design tool when they
applied it in addressing customer requirements in the design of engineering curricular. The
main focus of the study was the ‘Voice of Customer’ and its subsequent translation into
Substitute Quality Characteristics (SQCs) or means by which the organisation would respond
to each of the ‘Voice of Customer’ requirements. This was one of the most explicit
demonstrations in linking WHATs to HOWs and means to ends / objectives in education
using the QFD approach. Murgatroyd (1993) showed that QFD could also be applied in
Distance Education.
Grimes et al. (1994) considering students, academics and staff as customers used an
integrative Voice of the Customer to improve educational services. In this study, the student
is considered the primary customer and treating his voice together with those of the
academics (employees) constitutes the Six Sigma roadmap of DFSS (Ficalora and Cohen,
2009). Grimes’s approach was based on the Kano’s model, which helps in determining
priority areas for both customer and needs. Kano’s model (Kano et al., 1984) as in Figure 2.5
has been used over the years to classify customer needs into:
55
Type B (basic) attributes, or must be or expected
Type O (one-dimensional) attributes
Type E (excitement) attributes
Type I (indifferent) attributes
Type R (reverse) attributes.
The Kano Model can also be used to leverage quality attributes into both the ServQual and
the EduQual frameworks both of which are widely used for curriculum development and
programme implementation.
All Kano types of attributes are present in curricular design and implementation options and
management can, particularly under conditions of resources scarcity makes decisions of what
to emphasise and what to defer to future considerations. Based on market dynamics and
trending issues the Kano’s model can be a handy tool in determining organisational
actorhood.
56
In the same year, Jaraiedi and Rits (1994) applied QFD to the process of advising and
teaching where students played the key customer role. At the higher education level, the
teacher-lecturer as argued by Macfarlane (2012:25) has duties of being: a mentor, a guardian
of standards, an enabler of networking and resource for others and an ambassador for the
institution or discipline. The implications of focusing on ‘Voice of Employees’ in Six Sigma
lexicon is that HEIs would understand better how academics define themselves relative to
their job, the HEI and students, which is the main reason for the HEI being there. QFD poses
a challenge in this regard as it pull the professoriate from their hiding in the shadows to the
frontline of strategy formulation and implementation.
Clayton (1995) used QFD to ensure that the value of degree programmes is guaranteed at the
design, management, and delivery stages. These were the early indications of the potential of
QFD to shift the locus of quality assurance from extra-institutional structures to the centre
and fabric of the institution’s internal infrastructure. This research showed the inherent value
of QFD as a measurement, an assessment and evaluation tool. This served as the first
comprehensive signal of the intimate relationship between QFD and the Context-Input-
Process-Product (CIPP) model.
Ermer (1995) evaluated a programme's performance using a QFD model that considered
students, staff, and industry as customers. Ermer transgressed one confinement that had so far
existed: that QFD was just a methodology of doing work, and that the assessment and
evaluative power of QFD had not been explored in education; from whence educational
institutions could begin looking for ways of using QFD as both a methodology and an
evaluation tool.
Hillman and Plonka (1995) used QFD for the design of a curriculum. QFD was reported as
effective and supportive in the design process. The study showed how curriculum resources,
processes and outputs could be related to key customer needs. While Hillman and Plonka
heavily emphasise Voice of Customer and the Product Planning Matrix (PPM), the process
they described reveals lots of goal analysis, Target Values (expected learner outcomes) and
benchmarking (Competitive Satisfaction Performance).
57
Nilsson et al., (1995) successfully applied a QFD-based benchmark approach to evaluate a
Master of Business Administration (MBA) programme. In present day QFD, the process can
link Customer Satisfaction Performance and the sifting of Critical to Quality (CTQ) elements
of other programmes. The approach integrated QFD’s House of Quality matrices, which
made the effort all the more effective and efficient. In Taiwan in the same year, Chang and
Ku (1995) showed how QFD can be used to identify the improvement vectors in vocational
education. This is one further evidence of QFD’s use in Goal Setting and definition of
Improvement Ratio with the aim of aligning organisational processes to Customer
Satisfaction Performance.
Glen et al. (1998) studied the impact of QFD-based models and approaches in improving
educational processes and outcomes. QFD is used in process / business re-engineering, a
process that is helpful in realigning, and in modifying processes so that they become more
relevant. When well used, QFD should help in interface mapping and thus identification of
value-laden processes and those that are what Bevington and Samson (2012:208) define as
‘noise’; time and resources wasting and as work that adds no value to the organisation. Glen
et al., using QFD as a template, assessed and evaluated the value impact of context, inputs,
and throughputs to quality education (Stufflebeam and Shinkfield, 2007). This was yet
another attempt at using QFD as a programme evaluation model.
Eringa and Boer (1998) used a QFD-approach that hybrids Applied Service Blueprinting in
restructuring educational service processes that took students as primary customers. The key
exhibition of this work was demonstrating the versatility of QFD and its potential to work in
juxtaposition with other models. Ever then QFD has been used in the framework of Six
Sigma and Six Sigma has also been used inside QFD methodologies.
Franceschini and Terzago (1998) observed that QFD can be used in manufacturing and in
services industry, including education. QFD may have originated in the manufacturing
industry, but as a versatile tool it can still be applied in varied environments for service
creation. From then on, many researchers have begun drawing QFD lexicon into education,
58
with many others finding equivalent processes and alternatives. Franceschini developed what
is today known as ‘qualitometrics’ or measures of service quality.
Koura (1998) used QFD to determine student demands and requirements on how lecturers
delivered the curricular. This research touches what should be the core of quality assurance
work. Teaching and learning are the central tenets of an educational offering and if
educational institutions must save those who receive the learning then it must put them at the
centre of their strategies.
Lam and Zhao (1998) observed that QFD and Analytic Hierarchy Process (AHP) can be used
to improve teaching. AHP is a structured technique for organising and analysing
sophisticated decisions that need to factor goals, resources and strategic capabilities of the
organisation. Today QFD has increased in prominence as a decision-making tool. Owlia and
Aspinwall (1998) used QFD in improving quality in an engineering department. Their study
did what QFD is really for: delighting the customer by increasing Customer Satisfaction
Performance.
Bouchereau and Rowlands (2000) linked QFD to Fuzzi logic, to artificial neutral networks
and the Taguchi method to improve metrics for measuring organisation performance. This
study reiterated the value of QFD as a measurement technique. Furthermore it showed how
QFD can be used interactively with other techniques. A year later, Hwarng and Teo (2001)
applied QFD to the first three steps of the QFD’s House of Quality in higher education
institutions.
Ficalora and Cohen (2009) have mentioned that users can determine how much of QFD they
want to use, but as highlighted in Chapter 6, readers need to be extra critical in reading
research that claims to use QFD, but is actually using a single tool within QFD.
Inconsistencies between concepts and practice tend to create undesirable arguments on what
should change and what tools should be used in bringing in the desired goals. This fragments
effort that is otherwise critical for programme implementation and change in higher
education for example. Using a part component of a model is far different from using the
59
full-blown model and we are expected to be up front about that, lest they mislead the
readership. This concern is equally sounded by Al-Kassem (2013) who finds it misleading
about the main model or concept be it QFD or TQM.
Chang and Minj-Lu (2002) showed that QFD can be used in juxtaposition and interactively
with other models, techniques and tools. This point was made before and illustrates the
holistic nature of QFD. A point is reiterated that QFD is a third level-like ‘theory’ and its
analysis needs a pinch of systems thinking.
Sahney et al. (2003) studied the use of QFD in fostering productive relations between
industry and academia. QFD helps structure and manage collaborative relations by helping to
clarify intentions and expectations. Quality assurance work is being blamed for the lack of
integration between its otherwise important constituents. Salih et al., (2003) elaborate with
great success on how QFD was used in the design of a course on statistics. QFD is a ‘glue’
and communication tool that has great potential to bring people and their varied ideas
together. Yoshikawa et al. (2003) applied QFD to find strategies for easy learning and
mentoring in e-TQM projects. Using QFD’s roadmap Technology for Six Sigma (TFSS)
users are able to select appropriate technology and apply it to the creation, management and
delivery of value in their products and services.
Sahney et al. (2004a) used QFD and SERVQUAL in the gap analysis of quality education
and customer requirements in the education system. It was often observed that what higher
education institutions teach and what brought the student into higher education and the
particular institution are at variance. Chan et al. (2005) investigated the potential of QFD-
based curriculum design in the textile industry. The value of QFD in curriculum design
cannot be overemphasised. Aytac and Deniz (2005) used QFD in the revision of a
curriculum.
Bedi and Sharma (2006) investigated the use of QFD to develop case studies that focus on
curricular needs of MBA students. QFD was successful in processes (SQCs) regarding the
needs and expectations of students. Thakkar and Deshmukk (2006) examined the use of QFD
60
in educational competitive environments and found that QFD can improve the competitive
position of an institution. Indeed the ultimate deliverable in implementation is enhanced
competitive advantage. In the following year, Zheng and Pulli (2007) applied QFD to
improve the design of mobile services.
Ikiz et al. (2010) applied QFD to align educational structures to the Bologna Process. QFD
helped in achieving fitness for purpose of higher education offerings. External alignment is a
critical facet of QFD and the fact that it has proved to be able to establish it gives us the hope
that QFD can effectively work as an evaluation tool. In the same year, Mautsch et al. (2010)
reported that by using QFD one is able to achieve 'fitness for purpose' and enhance market-
orientation of academic programmes.
Thus QFD has been applied at the input, process and output levels of education, with
examples in staff and student recruitment, curriculum design, improving teaching and
assessment and evaluating post-qualification performance and satisfaction levels of alumni.
QFD has continued to be used in education to:
help in translating voices of stakeholders into educational products and services;
run Customer Satisfaction Performance with the idea of identifying performance gaps;
plan to close quality performance gaps between different elements of higher education;
link higher HEIs to their environments in search of external competences; and
assess and evaluate educational processes including management processes.
It will be of value that institutions attempting to implement QFD consult the available
literature and contextualise the many recommendations. The use and application of QFD in
African universities is not widely documented.
Thirdly, organisations need to realise that teamwork quality is the basis of QFD and thus
organisational politics that encourages workplace ‘gangsterism’ and yellow-banding is
counter-current to the spirit of QFD. Apart from teamwork quality issues, leadership need to
be at pulse with the happenings across the organisation, and leadership off the radar doesn’t
do well for QFD implementation (Hay and Fourie, 2002; Shar and Nair, 2014:147 / 152).
The fourth point is that the quintessence of QFD is gathering Voice of Customer, processing
it and satisfying customer requirements. Organisations that adopt QFD to deliver hay to the
horses will soon frustrate themselves when they turn to cost-benefit analysis.
Fifthly, QFD is a communication tool that should help management to glue people, processes
and issues throughout the organisation. When communication fails, QFD gets constrained.
Sixthly, higher education is a regulated business with multiple stakeholders who sometimes
have opposing expectations. Institutions that fail to negotiate amicable relations with
regulatory agents may find themselves concentrating on bickering rather than finding
synergies among them. Lastly, QFD is directed at helping organisations understand their
scoring on Customer Satisfaction Performance. More often than not, management impose
what they think of satisfaction onto the supply chain. Customers know their needs, and an
organisation needs to understand them and work out how to meet those needs.
2.13 Conclusion
The Theory of Constraint forms the theoretical framework of this study and basically affirms
62
that the success or failure of an enterprise is a complex function of the organisation’s ability
to manage both positive and negative constraints within both the formal and emergent
designs of the organisation. This understanding of the Theory of Constraint permeates all
discussions in chapters 2 to 6. Every often I highlight disconnects, hindrances and catalysers
of processes relating to quality assurance. The conceptual framework has been discussed
with the link between the five elements of the Strategy Focus Wheel and the 14 Best Practice
principles. QFD glues these aspects together and helps each of the strands and the principle
to optimise their performance. The three perspective of QFD are discussed and this
discussion challenges the appropriateness of the Harveyian definitions of quality in contexts
where QFD is followed. This new position challenges the traditional prerogative of HEIs
management to define quality disregarding the many disconnects their quality had with
student, Industry and society’s expectations. The Voice of the Customer becomes the new
central axis in defining quality because it is the epitome of what those who want the quality
say it should be. Discussing the growth and philosophy of QFD should help in understanding
how QFD has become increasingly relevant to the higher education sector over the years.
63
CHAPTER 3 – PROGRAMME QUALITY ASSURANCE IN THE PERSPECTIVE OF
QUALITY FUNCTION DEPLOYMENT
3.1 Introduction
Chapter 3 examines the potential link between the ‘Quality Function Deployment’ (QFD)
model and programme quality assurance (PQA), affirming that the link is possible, complex
and dynamic. It proceeds to compare and contrast QFD with nine quality assurance models
with the idea of establishing why QFD could be a first choice with most business schools. In
section 3.5 I examine the process of institutionalisation because the scope of adoption and
institutionalisation impacts the rendition of models. In section 3.6 I discuss inconsistencies in
quality assurance relating these to the nine models and examining how QFD could sooth the
gaps. Dimensions of quality assurance are examined and an evaluation of how QFD could
shift the locus of control and the prerogative to define ‘quality’ is suggested. Sections 3.9 and
3.10 highlight the gaps between stakeholder expectations of quality and the purpose of HEIs,
their behaviours and actorhood. In sections 3.13 and 3.14 the role of external quality
assurance mechanisms is examined, showing current deficiencies and how QFD could absorb
these roles into management agendas of higher education institutions rather than wait to
‘suffer’ them after their strategies are done. .
3.2 Towards a conceptual model of the relationship between QFD and PQA
In its infancy, QFD has been confined mainly in the manufacturing sector. Over the years,
researchers and practitioners have begun to notice the presence and usability of some QFD
tools in the services sector. Research into the increasing use of QFD has shown that QFD is
being conceptualised quite variably within and across the disciplines (Franceschini, 2002,
Ficalora and Cohen, 2009). For instance, those in economics tend to view and use QFD
slightly differently from the way those in medicine, banking and hospitality use it. Thus QFD
has grown to be a versatile methodology. In fact, Ficalora and Cohen (2009:7) say of QFD
that prospective users can adapt the model to fit their contexts and can make own decisions
as to how much of QFD they want to use. An attempt to relate QFD to PQA should ideally
show the direction of such relationship as well as a blow-by-blow analysis of the structural
and processual links between or among aspects of QFD and PQA. QFD illuminates PQA in a
64
number of ways, which can be summarised by arguing that QFD can serve as a strategy for
achieving PQA and ensuring sustainable superior quality performance. QFD offers a
systemic and systematic perspective to the conceptualisation, design and management of
quality and how the search for quality (‘qualitying’) should be done.
The endeavour of assuring quality is a ‘wicked’ and complex engagement where there are no
lineal cause-effect relationships. The absence of metrics whose validity and reliability is
unquestionable erodes stakeholder trust in claims of superior quality made by both Quality
Assurance agencies (QAAs) and HEIs (Harvey, 2006:187). Most reported successes are but
anecdotal. The other point is that some measures of quality are prone to distortion and / or
manipulation, as will be explained later under the Section on Measuring and Reporting. For
instance:
Can we actually measure quality by retention rate when there is no evidence that
students respond to midstream realisation of poor quality instruction or programme by
abandoning the programme?
Can we actually measure the quality of a programme by its popularity (enrolments)
when we have no evidence of the mediating effects of the market (for example market
signalling) or affordability (programme fees), or the amount and effect that market
signalling (marketing messages) impacted on the student and the sponsor?
Can we actually measure quality of a programme by its output rate without firsthand
understanding of its rigour, the psycho-pedagogical level of its content, scope of the
programme and the teaching methodologies being applied?
Can we actually measure the quality of a programme by an improvement in the
grades of the entrants without a critical analysis of changes in the grading regime of the
programmes or curriculum from which they are coming?
Can we develop satisfaction in quality assurance by saying there has been some
significant improvement since the last audit / visit when we have no means of
excluding management’s developed skills of concealing their weaknesses, or increasing
visibility of the few things they have been good at, and of whitewashing reports because
they now know what matters to the auditor?
65
These questions take us back to the need for a quality assurance system that is institution
based, and one that is within the genetic code of the institution. This is the case for a QFD
approach to programme quality assurance. QFD does not repudiate the need for external
quality management systems (EQMs). What it does is to take external quality management
mechanisms on board as inclusive of the organisation’s Strategic Quality Plan. Voices of the
external quality mechanisms are captured into the quality assurance process as Voice of
Business and Voice of Market. QFD blends Voice of Market and Voice of Business into a
Six Sigma roadmap – Marketing For Six Sigma (MFSS). Marketing for Six Sigma aims to
guarantee market orientation or fitness for purpose of the organisation’s products and
services. Blackmur (2010:67) and Singh (2010:193) say that for a long time, people have
been waiting for a model or family of models that can innovate and improve quality
assurance work. Some authors feel that quality assurance work is failing because of the
current structure between quality assurance agencies (QAAs) and those who do the bulk of
the work that is involved with quality, in the creation of value of education, its management
and delivery (Ramirez, 2013:126). Houston (2010) captures this disparity when he says that
academics and disciplines are largely excluded or exclude themselves from quality assurance
systems, and when they get involved it is superficial or they are involved in matters of
peripheral value. But still inclusion of people without inclusion of their points of view and
their experiences undermines their desires to contribute. Without genuine inclusion we lose
the opportunity for genuine commitment which would otherwise create a platform for what
Goleman and Senge (2014:12) called the ‘triple focus’. Triple focusing and what Senge et al
(2007 / 2012) call ‘presencing’ form the bedrock for a protracted and profound change.
The ideas of total inclusion have been within higher education debate for some time, as is
exemplified by Karmel (1990), Graham et al. (1995), and Jackson (1997), who advocated for
a shift in responsibility for quality assurance to the academics and the teaching-learning
processes. This shift is attainable in QFD since this model advocates for Voice of Employee
and the blending of policy, management and activities at the shopfloor level (Ficalora and
Cohen, 2009; Macfarlane, 2012). Such a shift would have an implication on the professional
identity of the lecturer, his curriculum orientation, and on the content and processes of
teaching and how the student learns. The speculative analysis is that the same course outline
in the hands of what Macfarlane (2012) classified as ‘academic citizens’ or ‘boundary
transgressors’ would be treated very differently and would presumably engage students along
very different vectors of intellectual growth. QFD does not only reiterate this mentality, but it
proffers the tools, methods and philosophy for deploying resources into the functions
(structures, processes, policy, etc.) that would create quality.
We can assume that the ideas have been taken and implemented by some, and these should
be shining examples of successful quality performance and quality assurance. Throughout
Chapter 2 I exhibited that thick quality is created by the disciplined deployment of the 14
management best practice principles across the organisation. Once the infrastructure of the
Strategy Focus Wheel and the 14 management best practice principles is in motion the
momentum for continuous improvement can be sustained. Continued failure of educational
programmes, projects and institutions implies that the discipline of strategic planning is not
completely mastered nor is the ability to create conditions for fructification of quality-based
models. QFD creates, in the long haul, a quality culture, which many feel is not present in
67
institutions that are low on quality performance (Shah, 2013; El-Khawas, 2014). The
implication here is that HEIs still have much to do in terms of having their own mass of skills
and in influences conditions in their institutions.
At section 2.4 the thesis alludes at the difficulty of drawing up metrics for each of these
constructs of quality. Traditional Quality Assurance mechanisms that focus on quality as
excellence, consistency and as value for money find it difficult to measure and report on
criteria derived from these constructs. Consequently related quality assurance efforts are
difficult to pursue, enforce and control for in an educational programme particularly in the
sense they are portrayed by Harvey and Green (1993). For example at Section 2.4 the thesis
discuss the difficulty in defining quality as excellence, consistency and value for money. If
there is ambiguity at the definitional level we assume there would be fragmentation at the
execution level and thus much difficult to draw concerted metrics for measuring how well
quality assurance is fairing.
A hermeneutic analysis of ‘FD’ (function deployment) alludes to ensuring that every process,
structure and interface in the system attains the highest standards desirable. Alignment and
68
integration are key to creation and delivery of high-quality products and services. This is in
consistency with the Theory of Constraint which considers incongruence between policy and
practice as a major threat to superior quality performance. Equally so, a hermeneutic analysis
of ‘PA’ (programme assurance) should imply that there are efforts to guarantee the presence
and functioning of structures, processes and interfaces with an eye to creating products and
services of high quality.
The quintessence of QFD in programme quality assurance lies in its potential to proffer tools
such as Correlation Matrices and Relationship Matrices. These tools facilitate in-depth
structure-function analysis, process analysis and interface mapping both of which are critical
in the identification of worthless activities that otherwise waste time and resources without a
commensurate addition on quality. Doing this is essentially assuring quality by deploying the
management best practice principles of alignment, being time-based and creating customer
value. Secondly, Correlation and Relationship matrices help in informing budget and
allocation decisions thereby improving the effectiveness and efficiency of the organisation.
These two characteristics give QFD an edge over other models: that of self reflexivity, of
being able to measure, assess and evaluate its own efficacy as a model. .
From these perspectives, achieving high quality can be seen as a result of improvements, of
change and innovation, all of which can be facilitated by QFD. Relating QFD and PQA poses
two challenges. The first relates to the theoretical compatibility of the QFD and PQA
constructs. To test for the possibility of linking QFD to programme quality assurance, it is
important to examine the following questions:
What are the key issues at the strategic, systems and operational levels of (programme)
quality assurance?
Does QFD inherently address all the key issues of (programme) quality assurance at the
strategic, systems and operational levels?
What are the measures and standards against which QFD can the evaluated and with what
intensity do these resemble measures and standards of quality in PQA?
69
At the strategic level, both QFD and programme quality assurance look at strategies,
resources, goal tailoring, and organisational structures that will best serve the requirements of
the stakeholder / customer. Quality assurance efforts that are not embedded upfront in the
strategic plans of the institution are likely to miscarry either because they don’t have case-
owners or resources or they are at variance with the programmes’s network of objectives.
While QFD defines the customer as the one who matters most for the survival, profitability
and market competitiveness of the enterprise this may not be the exact view in programme
quality assurance in its traditional sense. But with the advent and increasing prominence of
commoditisation, and marketisation of higher education in a deregulated market, PQA should
begin to take greater cognisance of these new factors. Complaints about Quality Assurance
seem to be about the outdatedness of their standards and points of view (Ramirez, 2013:127)
and their repudiation of the emerging business concerns (VoB), particularly for self-funding
programmes. We can therefore say that traditional quality assurance is retrogressive.
At the operational level, we can argue that the same measures used for testing the
effectiveness of QFD apply for PQA. Literature blames poor quality performance on a
number of issues. Firstly is the cleft between external quality mechanisms and internal
quality mechanisms (Vagnuer, 2008; Ramirez, 2013). Both custodians of the internal as well
as of the external quality management machinery have differing perceptions of what quality
is and how it obtains. Most external quality management systems are focused on control
mechanisms. Practitioners apparently abhor control and instead require more resources and
favourable policy regimes. Secondly are the intra-institutional fragmentations due to the
functional nature of organisational structures, reactivity and dysfunctional cultures which are
equally blamed for poor quality assurance in higher education. The divide between
management and academics in terms of what is quality and how it should be achieved
fragments HEIs and degenerate them into competitive and dysfunctional ivory towers.
QFD offers the tools for integration of efforts for quality assurance and the activities for
education design and delivery. QFD’s focus on cross-functional teamworking, on horizontal
integration and vertical alignment help managers and practitioners (academics) jointly assess
70
and evaluate their inputs in the perspective of Customer Satisfaction Performance.
Traditional quality assurance mechanisms look at products and services as having to meet
certain criteria. Normally these remain static over many years. This normally puts the Quality
Assurance agent at war with the more innovative HEIs. QFD instead looks at producing
products and services that chase the changing requirements and tastes of the stakeholder and
the customer. The QFD’s House of Quality Product Planning Matrix solves this tension.
Customer Satisfaction Performance and Competitive Satisfaction Performance are
predetermined by the provider and the traditional Quality Assurance agents. Normally these
are based on reactive and historic data, if not in the worst, embryonic transplants from other
contexts, mainly Western Europe or the USA. QFD uses both reactive and proactive data and
runs a blow-by-blow comparison of products and services with those on the market. More
examples could be given, but what matters now is to express the argument that at both the
strategic and operational levels QFD and programme quality assurance contain more than
meets the eye. The factors upon which the effectiveness of QFD can be measured are the
same factors that should define good quality:
Basing decisions on the basis of the voices of business, market, employees and customer
Collaborative working with openness of sharing perspectives and ideas
Meeting requirements of the state and its regulatory framework
Satisfying the needs of the industry, society and stakeholders
Designing products and services in accordance with needs of stakeholders
Correlating institutional goals and maintaining a healthy balance among them
Running institutional assessments and pinpointing performance gaps and measuring
Improvement Ratios
Defining absolute quality values and subordinating resourcing decisions to facts as
defined by objective targets.
HEIs can tout of adopting QFD, Business Process Reengineering, TQM, Six Sigma or any of
the ISO series models. But what matters is the adopting organisation’s ability to work the
model in ways that improve Customer Satisfaction Performance of the products and services.
It is the effects of the QFD on the totality of education delivery that the student, industry, and
society use to evaluate its value and meaning to them. Literature generally defines and
71
conceptualises this as perceived quality (Qp), which is a global evaluation of what the
customer feels he or she is getting from a product or service (Rust, et al., 1999; Franceschini,
2002).
The general outcry is that current quality assurance schemes are failing to find ways of
closing the gap / discrepancy between expected quality (Qa) and perceived quality (Qp)
(Franceschini, 2002). Expected quality is what the customer expects to get from a product or
service based on personal experiences, word of mouth and own imagination or expectations
(Rust, et al., 1999; Franceschini, 2002).
ΔQ(a-p) = Qa - Qp
Where: ΔQ(a-p) is the discrepancy between perceived quality and expected quality.
Qa is the expected quality.
Qp is the perceived quality.
The main reason for the existence of Quality Assurance agents is to close discrepancies
between expected and perceived qualities. This means working on the whole system of the
educational institution, from the mental framework of those at every level to the activities
that seek the smallest organisational goal. It means therefore that a change in our mode of
(re)thinking is far more important than any barrage of control measures. QFD powers what
Senge et al (2007) called the discipline of systems thinking which loops mental model with a
particular paradigm, set of means, objectives and outcomes. Figure 3.1 shows how QFD
provides this systems perspective. For instance a new paradigm (model or strategy) should
create a new set of means designed to accomplish a premeditated network of objectives (the
micro) that are derived from long-term goals (desired outcomes). Houston and Paewai
(2013:263) are of the opinion that ‘‘... attending to quality assurance from a critical systems
perspective provides a means of exploring its elements in a manner that does not ignore the
complex interactions that occur between them’’. Small-scale models that are particular on
isolated issues tend to ignore the environment from which the issues arise and its interactions
with other issues thus whatever solution is proposed it soon looks out of context and
proportion to the realities of the organisation. Let us plug Figure 3.1 into the discussion.
72
Figure 3.1: Using QFD to meet PQA in terms of needs of the various customers as synthesised from literature
73
Figure 3.1 indicates how the adoption of QFD will lead into the adoption of the means to
optimise processes and resources for the production of high-quality products and services
that meet all quality standards that are desired by stakeholders. It indicates the link between
the paradigm, the means, the objectives and the outcomes. Next I examine other quality
models. The focus is on comparing and contrasting the models so that we can identify gaps
and strengths from whence we can pronounce statements of assessment. Next we look at
some models that have populated the higher education landscape for some time now.
75
3.4.5 Utilisation-focused evaluation model
This model departs from the same premise as QFD: that evaluations should be based on the
utility and actual use of that which is being evaluated. Literature portrays this model as
backward-looking. This is what distinguishes it from QFD which is both backward looking
and forward-looking. Voice of the Customer is a key tool in each success-focused evaluation
as in QFD. Both can be used to evaluate how utile a lesson, course, programme and
curriculum has been. QFD can be used to predict how utile these will be and what needs be
done to optimise potential weak points. For both, evaluation is about the information needs of
the different stakeholder groups and how the findings shall be used (Patton, 1997).
Learning and acquiring skills that impact the large-scale quality performance of the
institution is a real necessity in QFD, as in programme quality assurance. As part of the
organisation’s desire to be out front, the building of strategic capabilities including human
resources competences should be treated specially (Dowling and Henderson, 2009:11).
Dowling and Henderson (2009:12) identify five categories of generic meta-competencies that
include: ‘‘self-awareness and reflection; self-management; communication competences;
leadership and influencing; and managing the performance of others’’. Each of these
categories has been ventilated in literature as important in quality assurance and in positively
influencing staff performance at the individual, group and team level. What is apparently
lacking is a solid research-based recommendation as to how to work the discipline of
organisational learning, thus little wonder why most HEIs are perennial fragmented
underperformers. Otherwise most HEIs are unclear about the difference between
organisational learning and a learning organisation. The first –organisational learning – is a
conscious longevity process of creating knowledge that improves total performance on value
77
creating and customer-satisfying goals; the latter is a nice-to-have-and-do thing not
necessarily connected to preconceived organisational goals. See Section 2.3 for a profound
discussion of the idea of organisational learning.
78
Each of the above models may be effective as small-scale, short-term tools for fixing
problematic issues. A more embracive and global model is better at dealing with quality
assurance issues systematically and systematically.
79
Colyvas and Powell (2006:346) classify institutionalisation into low, medium and complete
or full institutionalisation, the latter of which is said to involve substantially full exercise of
regulative, normative, technical and cultural-cognitive elements of the institution (Vukasovic,
2014:47). In terms of its scope, programme quality assurance work can embrace normative,
technical, regulative and cultural-cognitive aspects. The idea of university actor-hood is used
here to claim that universities don’t do any work, but the people within them do. Thus
institutional work in HEIs can be compressed into normative, regulative and cultural-
cognitive work. Figure 3.2 indicates the incremental nature of institutionalisation of a model.
Figure 3.2: The looped incremental relationship between elements of institutional work as
synthesised from literature
These areas are covered extensively within the stages of a QFD model and there is an array
of tools within those stages that help to enhance exploitation of each work area. Institutional
work on the cultural-cognitive elements of quality assurance is about building capabilities for
sustaining continuous improvement. It is also about learning and converting learning into
competitive advantage. The idea is that good behaviour should be institutionalised, but the
process of internalising ‘reason’ requires some training or other stimulus or catalyser, which
80
is where training, mentoring and coaching come in handy. In Chapter 2 both the theoretical
and conceptual framework highlighted the benefit in building strategic capabilities through
training, coaching, and mentoring as processes within the learning organisation and as part of
the behavioural objective approach; four-level model; the responsive evaluation model;
empowerment model and organisational learning models. These have been used over the
years in different contexts for quality assurance.
QFD’s House of Quality underscores the value of goal setting, of target values, and of
relating WHATs to HOWs through a number of matrices. Institutional work on cultural-
cognitive elements of quality assurance builds on institutional work on normative aspects of
quality assurance. Normative work normally involves the governance issues such as
democratisation of the decision-making processes, of institutional self-assessment, teamwork
quality, Customer Satisfaction Performance, and any work that smoothen institutional
management. Institutional work on the regulative aspects of quality assurance includes the
setup of a structure-structure, structure-function, and function-function relationship.
Alignment and integration of structures, functions and processes creates a strong base for
institutionalisation of a robust quality assurance infrastructure. Quality Assurance is more
meaningful when actors within quality assurance enjoy it rather than suffer it.
3.5.3 Managing the link between quality model (strategy) and organisational culture
The speed and depth of institutionalisation of a model should depend on the structure-
structure, structure-function and function-function relationships within the adopting
organisation. The interfaces among these and the content and processes of leadership also
influence the speed and depth of institutionalisation. It is critical therefore for management to
show sensitivity to the structure and mobility within the factor interdependency field,
including those that take place as the new model wedges into organisational life. Pearce and
Robinson (2009:397-401) discuss four ways or strategies of managing the strategy-culture
relationship. These include linking the strategy and culture to the organisation’s mission. The
second option is to maximise synergy between the strategy and the culture. The third bailout
kind of option is to manage around the culture of the organisation. Finally is the option of
reformulating either the strategy or the culture or both.
81
It is possible that an organisation can be working on one, two, three or all of these strategies
on different aspects of the organisation simultaneously. It is widely ventilated in management
and strategy literature that culture can affect the success of an organisation. For instance,
some aspects of the internal culture of academia have been seen by Kezar and Eckels (2002)
and Veiga et al. (2011) to constrain quality improvement initiatives.
Reformulating the strategy is a drawn-out exercise that includes transforming the ‘customs
and practices’ of the organisation. Care needs to be taken in engraining changes, particularly
with customs and practices that have become deeply engrained in the organisation’s
architecture. All the same, they have to be uprooted. If the cost-benefit analysis favours the
reformulation of the strategy or model, the organisation’s first choice may be to phase in the
new strategy or model bit-by-bit while using the Theory of Constraint and strategic risk
management tools to manage the organisation’s risk envelope.
Managing around the culture is the method whereby management introduces major
unfamiliar organisational changes as it paves the way for the new strategy. The lookout
points are to catalyse and invigorate those desired changes while weakening all forms of
resistance and weeding away growth inhibitors. Pearce and Robinson (2009:399) list a
number of ways of managing around the undesirable culture: creating separate entities; using
task forces, teams, programme coordinators; bringing in an outsider; outsourcing the service
or selling away the entity. These strategies have been variably tried in quality assurance
work.
Maximising synergy is a strategy that an organisation can adopt when it has to introduce
some desirable changes and is lucky to have the compatibility of the desired change and the
extant organisational culture. Organisational strategists and workforce will thus have the task
of synergising the desirable new and the existing desirables. The second task would be to
remove any roadblocks to the cementing of the link between the new and the existing.
Synergies can also be achieved between the organisation or parts thereof and externals, for
instance the quality assurance agencies. However, as reported in Harvey (2006:288), people
are more prone to react to suggested changes with hostility and disdain, particularly those
emanating from outside them.
83
3.6.2 Disconnect between purpose and metrics
Bevington and Samson (2012:208) define noise as:
‘‘... non-value-adding time, potentially distributed throughout the whole supply
chain, that is a direct consequence of not doing the strategically needed core
activity earlier in the process. Noise is the consequence of failure to do the job
once, right first time and therefore includes chasing, correcting and duplication.
Noise is always measured in resource hours’’.
Noise is measured in resource hours and another way of looking at it in monetory terms is to
total all hours lost by everyone who worked on the ‘noise-job’ plus the number of hours that
shall be lost in correcting the ‘noise-job’ by everyone involved plus the costs of ‘waiting’ for
the corrections. Rework (re)uses resources, time, and money gets wasted. From that we can
assume that noise in quality assurance work can come about when those who should
protagonise quality assurance fail to do one, two or so of the stages in QFD: institutional
analysis; strategy planning; engaging stakeholders; listening to regulatory requirements;
relating organisational competences to the needs of the stakeholders; seeking authentic
strategies of satisfying the stakeholder; learning effective and efficient ways from others;
setting goals and improvement ratios with the aim of continuously improving quality and its
accoutrements. There are numerous disconnects between processes, structures and mental
frames that should stand together in assuring quality in higher education. Stakeholders have
differing expectations of what universities should be teaching: should they focus on hard
content or on skills, and which content or skills to emphasise.
There is inordinate amount of disquiet on what should be the purpose of HEIs from whence
the purpose of quality assurance should be drawn (Altbach, 2012). There are also debates on
what quality is among quality assurers, as is the case among the actors (management and
academics and among academics themselves) in HEIs. Academics struggle with combining
rigour and relevance in the curriculum. Lodge and Bonsanguet (2014:3) highlight the
importance of addressing that link in research. What universities stand for is equally
equivocal among society. In this regard, Lodge and Bonsanguet (2014:3) are of the opinion
that ‘‘learning is core business for universities, though at times it might not be obvious that
this is the case’’. No wonder the world over is grappling with the increasing numbers of
‘garage universities’ (The World Bank, 2000:32), ‘pseudo-university’ (Altbach, 2001:8),
84
‘demand absorbing institutions’ (Altbach, 2009) and all sorts of ‘diploma and degree mills’
(Altbach et al., 2009) as well as ‘accreditation mills’ (Ezell, 2007; Altbach et al., 2009:60).
Even for the recognised institutions, judging by a selection of objective measures, there is
little to build confidence that HEIs are delivering robust, value-adding benefits for society,
industry and their students (Shermis, 2008; Arum and Roksa, 2011; Lodge and Bonsanguet,
2014:4). This suspicion comes from research and experiences that confirm such perennial ills
in the higher education sector, which Bamber and Anderson (2012) describe as the absence
of theory that guides the evaluation of the complex instructional process. The proliferation,
popularity and disdain after forced closer by quality assurance agencies of some of the so-
called informal providers should send us re-examining our metrics of quality.
The situation on the ground is one where standards and measures are doubted. In fact, the
section for quality assurance in the higher education landscape is littered with tools and
techniques that don’t measure what we think we are measuring. QFD may provide a new and
more relevant way of assessing the HEIs than the ‘‘... poorly defined and confused
performance indicators’’ currently in use (Lodge and Bonsanguet, 2014:4). With the
explosion of information technology, the number and nature of tools available to those at the
terminals of technology has exploded. Most HEIs have resorted to ‘cut and paste’, thus
multiplying the nature of generic tools and instruments for assessing quality learning.
As instruments are multiplying, apparently the cleft between instrument-to-purpose fit have
widened. Most collect data from students. This has little strategic value. What has strategic
value is data about the student so that action about the student is taken. Tracer studies have
shown the ubiquity and longevity of such deficiencies in the higher education sector (Bok,
2006; Borden and Young, 2008). One of the cumulative effects of such disconnects is the
strategic drift of the university from its supposedly core business of learning and the ways in
which they are evaluated and legitimised (Lodge and Bonsanguet, 2014:5). There is little
wonder why literature finds university rankings of no value to the core of teaching and
learning. If we should assume that the core of quality is learning, then equally learning
qualifies as the chief or key constraint in the instructional relationship. By this thread, and the
discussion in Chapter 2, everything else then would need to be subordinated to the learning
85
deliverable(s). The most ventilated constraint is our way of thinking or responding to issues
around us and we referred to this in Chapter 2 as ‘mental models’. The thesis’s argument is
that the way we think and conceptualise quality and quality assurance, how they can be
gotten and how they matter to every one of us should determine our success in the pursuit of
quality and quality assurance. Coming back to the question of mental models and systems
thinking we observe that stakeholders are at variance about the conceptualisation of learning.
To exacerbate issues, most HEIs are reported to be using proxy measures of the learning
experienced by students, thus adding strength to the web of distrust about the actual value
that HEIs are adding to different stakeholder groups.
Measuring instruments are condemned for their lack of validity and reliability (Porter, 2011),
lack of usability (Harvey, 2003; Richardson, 2009) and a host of other nuances (Lodge and
Bonsanguet, 2014:7). Earlier, Astin (1991), Heywood (2000) and Race (2006) had pointed
out that most instruments don’t measure what we want measured or just don’t measure what
they should. Boud and Falchikov (1989), McCabe (2010) and Bowman (2011) found that
students are often at sixes and sevens about value added to their knowledge, attitude, skills
set, understanding, behaviour and belief system (KASUB) by their scholarship. To enhance
their accountability profile, HEIs are shifting toward the use of objective and comparable
measures of student learning outcomes (Marginson, 2009). With increased quality literacy
levels, more and more people are able to decipher inconsistencies between instruments and
what their users claim to be getting by using them.
Figure 3.3: Accountability for, to, and how as synthesised from literature
The question of accountability in QFD contexts takes very different dimensions to the ones
ventilated in literature and is exemplified in the diagram above. In QFD the target of
accountability is the customer. The purpose of the organisation in QFD is to meet the ever-
changing tastes and preferences of the customer. The customer is the key source of revenue
and deserves the greatest amount of accountability. By running Voice of the Customer the
organisation’s aim is not to hoodwink the customer into thinking that he is king. The idea in
Voice of the Customer is to see how the organisation can be made more customer-oriented
and by focusing more and more on continually improving his satisfaction the organisation is
paying back for the time and effort lost in responding to Voice of the Customer surveys.
Whilst accountability in traditional higher education is to the government and the local power
87
structures, in QFD it is all about giving the customer the greatest in terms of products and
services that exceed his expectations. Up to this point we have loosely referred to various
dimensions of quality assurance.
A magnitude in quality assurance can be the number and diversity of the criteria that are used
by internal and external quality assurance agencies. Quality assurance schemes may range
from cosmetic, compliance-focused to improvement-oriented ones. The direction of quality
assurance may as well be externally motivated or endogenous. The direction can be positive
and that is when the idea behind the effort is to increase or enhance what is already present.
The vector of quality assurance may be considered negative when it is about reducing and
doing less of something. In QFD’s Kano model this corresponds to Type R (reverse)
attributes. In one of the stages of QFD, goal setting and improvement ratios are discussed and
it is not uncommon to find management opting to ignore certain issues of value so that
resources are freed so that they can be used for doing more of another type of attributes. In
non-innovative visibility-seeking organisations emphasis on Type E (excitement) attributes is
common. It is not unimportant to ‘excite’ one’s customers by adding a number of ‘wow’
effects in the programme. The question is about the morality dimension of quality when it
comes to having high quality facilities and cars when the core of teaching and learning is
suffering without resources. At the international level, quality assurance is about finding
common understanding of factors for qualification recognition. It is also about learning from
one another. At the national level, the focus of quality assurance is the national development
priorities. At the discipline level quality assurance covers parity of standards based on
88
qualification descriptors and a system of subject reviews. At the institutional level, quality
assurance is about the presence of a risk management infrastructure that encourages quality
enhancers whilst weeding away threats to quality.
At the programme level, quality assurance’s focus is on how the individual subjects deal with
the development of knowledge, attitudes, skills set, understanding and behaviour of
graduates. At the subject level, quality assurance is about the rigour, relevance and fitness for
purpose of the instructional relationship. At the lecturer level, quality assurance is about the
lecturer’s credentials, psycho-pedagogical adaptability and content mastery. At the process
level, quality assurance is focused on a processual relationship in which the student learns
unlimitedly. Whatever its scope, quality assurance should seek to integrate improvement
work and an orientation for innovation in a never-ending manner. Figure 3.4 exhibite this
desirable.
Whilst QFD has clear cognisance of the two orientations, traditional Quality Assurance
models tend to encourage only the improvement aspect of quality assurance. Aytac and
Deniz (2005:512) conceive that methodologies like QFD can be used as a guiding tool for
improving educational activities at every level of curricular activity and by any player with a
stake in educational quality. National quality agencies have also sought knowledge,
competences and advices on how to improve and innovate from across their country
boundaries. This is exemplified below.
89
3.8 Global and international efforts at quality assurance
The value of cross-border collaboration in quality assurance (QA) is increasingly being
recognised globally and in Africa (Materu, 2007:37) with more and more quality assurance
agencies increasing their operations across the borders of their national origins. Specialised
programme accrediting bodies that are now international players include the US-based
Association to Advance Collegiate Schools of Business (AACSB) and the Accreditation
Board for Engineering and Technology (ABET). The European Quality Improvement System
(EQUIS) (OECD, 2004) and the International Network for QAA in HE (INQAAHE), which
was established in 1991 to collect and disseminate information on current theory and
developing practice in the assessment, improvement and maintenance of quality in HE
(Materu, 2007:37) has become international players as well.
European Union nations, in 1999, launched the Bologna Process, and invested in it in the
hope that it would strengthen European cooperation in QA. The Asia-Pacific region launched
the Asia-Pacific Quality Network (APQN) with a mandate to cohere quality assurance efforts
in the region. The Latin American region launched the RIACES to spearhead quality
assurance initiatives across the region. Within Africa, rather fragmented efforts have been
shown, yet with a common interest in assuring the superior quality of higher education. The
only body that has a continental focus is the Association of African Universities.
The other institutions are fundamentally regional: Southern African Regional Universities
Association (SARUA) and the Higher Education Quality Initiative (HEQMISA) for SADC
countries; Inter-University Council of East Africa (IUC-EA) for East African nations;
Counseil Africain et Malgache pour I’Ensignement Superior (CAMES) for francophone
African nations; North Africa and the Middle East-Arab Quality Assurance Network for
Higher Education (ANQAHE) for Arab-African nations and Arabic nations of the Middle
East. If all these organisations have a shared vision, about a similar customer, and similar
goals then teamworking and the application of QFD to their work should be feasible and very
rewarding. Why these organisations continue to fail to find synergies and collaboration
among them is arresting. It is interesting that they continue to be protagonised by academics,
90
serving as employees. Thus however discussions may happen, it remains doubtful how they
represent positions of other absent stakeholders to higher education.
Harvey (2002) however laments that quality assurance processes pay little attention to
educational theory, educational processes, or student learning and are not aligned to what is
generally accepted as purposes for higher education. The United Kingdom’s quality
assurance agency sees master’s degrees as “designed to fulfil a range of purposes” reflecting
“both the desires and ambitions of students and the traditions and needs of particular
disciplines and professions” (QAA, 2010:4) and “prepare students for the next stage in their
career” (p. 13) and “to contribute toward research in the discipline” (p.11). Jabnoun
(2009:416) proclaims that HEIs are a source of national pride and ‘‘play a pivotal role in
economic competitiveness and the sustainability of economic growth’’.
A World Bank Report claims that with good higher education systems nations are better able
to deal with health, development, engineering, and socio-political issues. The document
91
further states that higher education plays a catch-up role particularly in creating the critical
mass of skills in the manufacturing and services sectors (The World Bank Report, 2000:17).
Bloom et al (2005) and Hayward (2006:6) report that higher education may help Africa
broaden its manufacturing base, promote further technological catch-up and improve its
nations’ abilities to maximise economic output. Houston (2010:178) and Williams (2011:1)
argue that the purpose of higher education is to promote learning through core functions of
teaching and research. They further assert that quality initiatives should therefore result in the
improvement of teaching and research. However other literature show that little has changed
since Harvey (2002) decried that quality assurance processes apparently paid little attention
to educational processes, educational theory or student learning. Houston (2010:179) says
that higher education should actually empower and encourage lifelong learning in ways that
benefit society.
Most literature concur on the purpose of higher education. Based on a wide ranging study the
QAA (2010) discovered that master’s degrees may be designed with more than one purpose
in mind, among these being to enable master’s graduates to:
focus on a particular aspect of a broader subject area in which they have prior knowledge
or experience through previous study or employment;
focus on a particular subject area or field of study in greater depth than they encountered
during the course of previous study and experience;
learn how to conduct research often linked to a particular discipline or field of study and
thus contribute to the discipline;
understand a research project on a topic within the area of interest that makes up the
majority of the overall assessment;
specialise or become more highly specialised in an area of employment or practice
related to a particular profession; and
prepare themselves for the next stage in their careers, whether pursuing further research
or entering employment of different kinds.
What then should be the attributes that we would expect to see in a master’s degree graduate
if we consider the perspective of the ideal purpose of higher education and the ideal second
92
cycle curriculum? To what extent is the curriculum fit for purpose? It can be concluded that it
is important for both arms of quality assurance to focus on the design and delivery of
curricular that will ensure quality higher education. A lot of research has focused on the
development of modes of focusing on some of these purposes.
3.10 Philosophy and methodology for curriculum development, quality assurance and the
idea of ideal master’s graduate
Graduates of master’s degree programmes are expected to have a certain level of skill that
should help them to be employable, to be employers and protagonists of many socially
responsive initiatives. The methodology of instruction must therefore align with the
expectations of the ideal master’s graduate. The philosophy of education as well as of quality
assurance must then ensure the presence of an appropriate methodology of quality assurance
as well as of teaching and learning. Figure 3.5 show the proposed relationship. This kind of
link is apparently lacking in national and institutional quality assurance structures despite
being widely present in most literature (Bowden and Marton, 1998, Coates, 2010). This
shows the difficulty of making the proposition work which may be because the proposition is
excessively abstract.
The traditional approach to the relationship between ideal graduate and philosophy of
education is that the philosophy precedes and determines the quality of the programme
graduate. What if the philosophy is erroneously premised? A philosophy may become
dominant because it is protagonised by a group that has absolute power within a discipline or
society. Philosophies should be as good as they are collaboratively built by those affected by
them. This is the basis of Voice of the Customer in QFD.
93
Graduates of master’s degrees are expected to have a bundle of subject-specific attributes as
well as generic attributes. In a general sense, master’s graduates would be considered to have
the abilities to take responsibility and use initiative, and to offer creative and innovative
solutions to all sorts of problems. They should be able to initiate and take decisions in
challenging situations and develop themselves into meta-cognitive learners. Master’s
graduates should be able to communicate effectively and adapt strategically with efficiency
and in ways that are effective. Further to the repertoire of generic skills, master’s degree
graduates are looked upon to exhibit subject-specific attributes that include profound
knowledge and understanding of the discipline. Such profound mastery of the discipline
should be rooted in lifelong learning, awareness of the currency, evolutionary trends in the
discipline and other developments around the knowledge areas.
Students graduating with a master’s degree should be able to complete a research project by
way of critical literature review or by empirical methodology. This requirement calls for
proactive behaviour, analytic abilities, synthetic abilities, teaming abilities, communication
abilities and abilities to solve problems. Sahlberg (2006) strongly advise on the need to
abandon the old fact-based curricular and adopt curricular that place focus on developing
thinking skills, interpersonal skills and creativity. Other models challenge higher education
programmes to develop strategic skills (Scott, 2008).
Alex Scott (2008:38) postulates that ‘strategic thinking’ arises from an in-depth command of
various academic subjects that cover management, financial, economic and analytical issues.
From such a foundation of knowledge, the graduate should be able to identify models that
suit his circumstances as well as use prescribed models to navigate through workplace and
life issues. Sahlberg further affirms that one can actually be considered a strategic thinker
when one’s synthesis and evaluation skills are sharply developed. Other skills identified as
critical for life and employability include emotional intelligence and social intelligence. In
fact, Lowden et al. (2011, p.4) talk of literature that has developed taxonomy of skills. These
skills are classified as: ‘‘core skills; key skills; common skills; transferable skills; essential
skills; functional skills; skills for life; generic skills and enterprise skills’’. To what extent is
the work of the various quality assurers guided by these taxonomies of employability skills?
94
Owlia and Aspinwall (1998) rank-ordered factors they felt are determinants of the overall
quality of graduates. These are the quality of delivery and management of programmes of
study; recruitment, appraisal and development of the staff; rigour with which study
programmes are designed; guidance, counselling and support given to students; quality of
admission criteria; service support of study programmes and assessment of students. Apart
from this identification there have not been robust ways of improving the focus on these
aspects with regard to the higher education strategic planning and Quality Assurance agents’
formal designs. An analysis of literature show that we are awash with ideas about quality
assurance but not as much with the strategies of how the ideas can be made to work.
One of the generic purposes of quality assurance mechanisms is to inform the public on how
well higher education institutions (HEIs) are meeting public expectations and what they
declare either in good faith or as marketing gimmicks. Because of the heterogeneity of
international economics, politics, markets, cultures, etc the term quality has become more
and more contextual and stakeholder dependent. In 2006, the Vice President of the European
Association for QA in Higher Education said that because of the elasticity of the term
‘quality’ Quality Assurance mechanisms must be equally flexible and follow the trend if they
are to be relevant and of service at both the national and international level.
To improve their relevance, Quality Assurance systems must measure both the fitness of
purpose (focus on process) and the outcomes of higher education (focus on the graduate).
Nunan and Calvert (1992) argue that where measurement of quality focuses on the student as
95
a product of education, quality is seen as value added by the process of education. They
further argue that when the emphasis is on management of quality, the focus shifts to
strategies for achieving and improving quality (Nunan and Calvert, 1992). Both
transformation of the student and the institutionalisation of quality improvement strategies
must be seen as interweaving facets that have a mutual enhancement effect. A quality
assurance perspective that encourages and values diversity above certain thresholds proffers
incentives to institutions seeking to brand themselves with new products, services and
processes. While there is abundant agreement that the process level of educational activity
defines the core of quality, the reality is that there is also the level of education that is most
difficult to measure and to monitor. Teaching and learning happen far from the control and
monitoring activities of management and of the quality assurer. Poole (2010:6) feels that the
narrowness of certain definitions of quality may misinform quality assurance work. Further
to this observation, Rowley (1997:9) thinks that the definitional ambiguity on ‘quality’ would
potentially confuse any efforts to quality assurance.
The public must be able to see evidence of quality in the service. An effective quality
assurance system would need to be a detailed and thorough cyclical process rather than a
seriatim of brief and superficial runs of checklists. Due to the increasing diversity of the
student market and delivery systems in higher education, the quality assurance measures
need to be sufficiently flexible and diverse. Indeed, quality assurance systems must promote
96
diversity. Quality assurances are both inward and outward looking and their deliverables
must be disseminated in terms that are understood by a lay audience.
To be of both corrective and developmental value, quality assurance systems must address
the issue of standards and be conducted at both the programme and subject level. In this
epoch of globalisation, quality assurance mechanisms should contain international
comparative measures and should be conducted by both national and international peers. A
huge concern in higher education today is how to translate employability skills and industry
requirements into the curriculum. Further, the credentials of those in quality assurance must
be of high probity.
Quality assurance systems in Africa are still in their embryonic stages of development and
are experiencing numerous constraints that include the human capacity in terms of expertise
and availability for quality assurance projects and jobs (Materu, 2007). It goes without doubt
that the credentials of the assurer of quality in higher education is as important as should be
the quality assurance infrastructure itself. There is lack of competence and capacity
development for those in quality assurance assignments at the accrediting agencies, at
institutions and within peer review teams (Materu, 2007; Massaro, 2010). Furthermore, there
are massive problems at the accrediting agencies and at HEIs to collect and process the data,
97
information and self-analysis needed for effective self-studies (Materu, 2007:49). The quality
of those at the forefront of assurance mechanisms has an effect on the quality of response
their organisations receive from those for which and upon whom quality assurance is done.
Materu (2007:65) alludes that the integrity, credibility and legitimacy of the work done by
quality assurance agencies depend on the credentials of the professional staff leading the
quality assurance process.
Professionals at the national quality assurance level need skills for system conceptualisation,
the development of methodologies and for rolling out of quality assurance systems.
Effectively, professionals in quality assurance are boundary spanners, connecting with an
unlimited galaxy of constituents. They need strong negotiation skills. Quality assurance
strategists need adequate management capacity; such that midstream changes of directions,
scope and measures are avoided as much as possible. Continuous capacity building is
necessary as a way of ensuring that the quality assurance system is oiled and smooth flowing.
In fact, the quintessence of quality is in the processes that are designed to bring about and to
sustain it, and these evolve over time.
3.13.2 Licensing
It is not uncommon with some institutions of higher learning to take advantage of most
stakeholders’ confusion regarding licensing, accreditation and quality assurance. Licensing
can be a once-off or phased process that culminates in an institution being granted authority
to launch products and / or services within some given framework (Materu, 2007:4).
Licenses are normally granted by ministers of higher education and the requirements for
quality that normally go with licensing are not stringent enough.
The involvement of a minister of higher education in the accreditation decision has an ‘eye
washing’ effect, making things look good and of higher standards. While ministers may
represent society, responsibly in other cases, it is difficult to do so with a process as complex
as quality assurance, taking into account that ministers have other roles. Further to that,
ministerial positions are political in nature and often subject to high turnover (Materu,
2007:22). These factors debilitate the credibility and trust that the system of assuring quality
could court from the public. An infrastructure for external quality management consists of
processes such as accreditation, licensing, site visits, inspections, etc.
3.13.3 Accreditation
3.13.3.1 Introduction
Accreditation is a process of validating the infrastructure, processes, products and services of
an educational institution for being above a set threshold. In most cases the standards of
accreditation are set by some expert group. The important thing in accreditation is pitching a
link between accreditation standards, stakeholder concept of quality and the accredited
institution’s buy-in to the process and its deliverables. Accreditation should be a more
99
disciplined and rigorous process directed at ensuring conditions for effecting quality
assurance.
Accreditation pursues some definite objectives with these guiding principles. On passing the
judgment to accredit an institution or programme, the most common practice is to seek
evidence of successful practice in a number of operations / criteria. What is likely to be
overlooked is how deeply the various criteria are interweaving and how the programme or
institution subordinates the rest of the criteria to teaching-and-learning. It is the teaching-and-
learning that transforms the students and makes them fit for a purpose. The picture that
should guide quality assurance is one that revolves around teaching-and-learning.
101
3.13.3.4 Benefits of accreditation
Materu (2007:29) observed that there is a growing body of evidence that accreditation is
beginning to take a prominent role in institutions that are seriously mindful of quality
assurance. Accreditation can also help to check out entrepreneurs fuelling the willy-nilly
proliferation of demand-absorption institutions from having a field day by milling low-
quality programmes (Altbach et al., 2009:4). In conditions of ‘promotion / employment-by-
qualification level’ – individuals scrambling for the nearest qualification that give them an
employment or promotion opportunity – higher education providers will make huge profits
(Altbach et al., 2009:4). Altbach et al. (2009) raise the fear that unchecked multiplication of
pseudo universities has a lethal effect on ‘true’ universities, a concern once raised by Akerlof
(1970) when he described ‘lemons’ as poor quality products and services. He went further to
raise alarm about the potential of ‘lemon’ qualifications to destroy the market of truer and
rigorous qualifications. QFD has the potential to draw the requirements by external agencies
into the strategy and product planning matrices with consequential improvement of the
programme quality and fitness for purpose of the graduates.
Accreditation has the effect of reducing variances between domestic programmes and those
offered abroad. Standards that are clear and more convincing are more likely to encourage
institutions to adopt them. The aptitudes of entering students have an effect on the quality of
teaching and learning and the higher these aptitudes are the more likely teaching can be kept
more rigorous and of superior quality. Physical inspection and visitation of institutions puts
pressure on higher education managements to improve the quality of their facilities (Materu,
2007:29). Continued deterioration of university buildings, classrooms and other resources has
been widely observed in many African universities (Garwe, 2007).
In efforts to keep high standards, accreditation boards impress on minimum qualifications for
faculty and regulations of promotions and staff recruitment procedures. These antecedents
are designed to bottleneck entry and promotion of sub-standard senior staff in institutions of
higher learning. Accrediting boards and some research funders encourage the use of
appropriately qualified part-timers so that the full-time faculty can have opportunities for
self-upgrading and undertaking research. Accreditation boards undertake crucial activities
102
like recognition of prior learning and the evaluation and validation of credentials obtained
from extraterritorial institutions. This helps to close the clefts between domestic and
internationally recognised qualifications. The regulation of teaching resources through setting
minimal standards of human resources competences and technical resources help university
managements keep conditions above threshold levels. Most African countries are shifting
toward institutional accreditation rather than programme accreditation with the danger that
once an institution is duly accredited, the process itself masks deficiencies at micro
(programme, course) levels, like a programme. While institutional rather than multiple
programme accreditations is cost saving, it needs to be augmented by a rigorous process of
reviews and self-assessments aimed at meeting the best levels of accreditable quality.
Accreditation is normally phased into: self-assessment; peer review; site visits, and a written
report. The assessments for accreditation focus on judgments about capacity, quality,
outcomes, and the need for improvement. Materu (2007:19) reports that in Cameroon, Kenya
and Ghana public universities are ‘accreditated’ de jure. De jure accreditation fast-tracks the
process, thereby bypassing the rather healthy stages of self-study, peer review and site visit.
With guaranteed demand and de jure accreditation earned by virtue of their being public
institutions, public higher education institutions are likely to allow trade-offs in quality to
accommodate the social demand for access and to offset the effects of reduced funding from
government instead of confronting the challenges of searching for alternative solutions
(Materu, 2007:31).
Reynolds (1990) observed that academics welcomed the external examiner system,
particularly in foregrounding arrangements for quality assurance. Institutions have found the
104
process of external examiners useful, benefiting from their external views and comments on
current and planned provisions (Materu, 2007:19). At some African universities external
examiners are recruited from regional universities, but other institutions have completely
stopped using external examiners due to the lack of finances (Materu, 2007:19). Variably
external examiners are given powers over final marks – a move that lends credibility and
legitimacy to the final grades (Materu, 2007:17). However there are numerous difficulties
that have been experienced with the external examination system.
The QAA (2009:20) reports on an experience where “senior institutional managers were
instructing their staff to tell external examiners that their institutions needed more first class
honours degrees in order to improve their standing in the (unofficial) league tables published
by national newspapers”. Other disappointing cases of meddling with external examiner
functions involve institutional managers overturning external examiner decisions despite the
manager having no expertise in the subject area.
Other observations involved external examiners being pressured to alter reports (QAA,
2009). The immorality of both some external examiners and some managers has left the
external examining system in a dishevelled state, becoming a fig leaf that cannot guarantee
academic standards any longer. However, where such irregularities have not been
105
experienced, the confidence is high among faculty and administrator that external examiners
play a strategic role in ensuring fairness in the assessment and comparability of institutions
and course standards. External examiner contribution in cross-institutional comparisons will
be constrained by the large diversity of institutions and their interaction with the host
institution is time-limited. The lack of expertise in the many courses on offer is another
limiting factor. Some African universities have done away with using external examiners
however in cases where they continue to be engaged, they are domestically recruited and
their stay periods have been cut to save costs (Materu, 2007:16). This means that their
workloads have increased and this may affect their efficiency.
Due to institutional massification, external examiners may find it impossible to go through all
candidates' scripts. The quality of faculty is falling and this lowers the quality of external
examiners that are available, thus too lowering the quality of external examination. Some
difficulties with external examining arise because the recruitment is not as open, transparent
and informed by the future roles to be run by the external examiner. In this light much can be
done to improve renditions from the external examining system. Notwithstanding the
importance of increasing the importance of internal quality assurance remains critical and
QFD provides that vehicle.
Entry into an ideogora could be screened, say, by institutions and quality assurance agencies
that have experience with the particular academic. It is also very important that management
do not interfere in the work of external examiners so as to assure undue influence and the
106
lack of duress in report writing. Shuib et al. (2007:1) are of the opinion that the use of
external examiners, movement of academics around institutions locally and abroad,
involvement of professional associations, and allocation of research grants by competitive
assessment have had a continuous effect on the exchange of information and the maintenance
of high academic standards. Whilst an institution’s response to rankings is optional and if
done can be quite unique the possibility that lecturers and students in their individuality may
be upset or inspired by some aspects of rankings cannot be denied. In this way ranking exert
a latent micro-level force even on organisations that don’t subscribe to the rankings
regiments.
Most ranking schemes exclude qualitative aspects of education, yet much that shows good
education can only be understood in qualitative terms. It is therefore important that before
ranking can be used the potential user departs from a complete understanding of how the
ranking framework of criteria is designed and the purpose it was designed to serve. A ranking
scheme that is designed to measure research activity may not be of much help to nascent
universities that are struggling with cash and quality staff. It sounds reasonable then that
when HEIs’ managements decide to draw lessons from rankings they should look for ranking
types that fit their contexts and level of development.
Some authors arrange the QFD’s House of Quality into four phases: product planning;
product design; process planning, and process control. One thing that stands out saliently is
the lock-step nature of the four stages. This is important in guaranteeing that no work is done
outside the scope of any premeditated objective. This principle of QFD helps in saving
resources and time, aligning resources to goals and giving each person more time to focus on
their own productivity. These issues are discussed with profundity under the theoretical
perspective and conceptual framework in Chapter 2. Other 4-phase QFD models that
emphasise different structures, processes and contents are not unknown on the market. For
instance as in Figure 3.6 one such model emphasises the phases of: House of Quality; parts
development; process planning, and production planning. In the first step: House of Quality,
110
the Voice of Customer is processed into customer attributes (WHATs / requirement /
objective).
Each customer attribute is then linked to a design characteristic (SQC / HOW / Response) or
a means by which it will be achieved most effectively and efficiently. The proceeding stage
is Parts Development. In ‘parts development’ the design characteristics takes up the value of
the WHAT or OBJECTIVE and a new set of responses or HOWs which are the features of
the products and services become the means of achieving the Substitute Quality
Characteristics. In the third stage: Process Planning the Parts characteristics assume the
position of objectives that should be attained by designing key process operations. A Six
Sigma Process Design roadmap can be used. In this way wastage is controlled as well as a
best fit between the desired features and the process of creating them is achieved. The last
stage of this four-phase QFD methodology is the Production Planning stage. Here achieving
the key process operations become the objectives of this stage. The matrix matches each of
these objectives to the production requirements. This is a shorthand model of QFD. In
Chapter 2 I mentioned that QFD is a versatile tool and adopters can make choices as to what
of it and how much they would like to use.
111
The point of departure is the voices of the various stakeholders / customers and these are
arrayed each against the strategies that will be used to accomplish each need expressed by the
customer / stakeholder. In the next stage, the strategy itself becomes the goal against which
ways / means of achieving it is conceptualised. This goal-means relationship goes on through
the other stages. The idea is to see issues of greater value and to make sure the organisation
invests only in what is required and what it will use. Some authors and practitioners show the
House of Quality as consisting of five key areas of activity: technical descriptors; customer
satisfaction performance; institutional assessment; Voice of Customer and relationship
matrices. Six-roomed House of Quality is not uncommon in literature. One such six-roomed
House of Quality is shown in Figure 3.7 below.
In fgure 3.7 the first room would be the Voice of the Customer. Apparently the forms of
House of Quality used within QFD recognise the importance of premising quality assurance
on Voice of the Customer. The second room is the voice of the designer or the people who
will finally shape the goods and services. The third room represents the relationship matrix,
112
the forth the technical competitive position and target values, the fifth the market competitive
position and the roof the design interactions.
Ficalora and Cohen (2009) lament that those who claim QFD does not work are not creative
enough to contextualise the QFD lexicon nor some of its tools. It is important that those
reading about or using QFD be creative and imaginative enough to think of equivalent terms
or processes. For instance in manufacturing QFD, we talk of engineers whose equivalence in
educational QFD are the professoriate. Both deal with the application of social, economic,
scientific and practical knowledge as they research, invent, design, develop and improve
structures, processes and new modes of thinking. Looked at in a rigid and mechanistic way
any model would never work for anything. There are no contradictions in these approaches,
practitioners adopt most convenient constructs of House of Quality, but this may pose some
pragmatic dangers.
QFD uses numerous maps, charts, tables, matrices, statistics, flowcharts and other tools at
many of its stages to help users gather and process data so that they reach useful decisions.
Stages in the QFD approach / methodologies are variably referred as tools too. The QFD
flow in Figure 3.8 below is not much different from the contents of either the five or six-
roomed House of Quality shown above. The diagram below is more elaborate, showing
stages where some are subsumed in other stages of the former two. A strength of the more
elaborate diagram is that it shows the side steps of ‘Analyse results thus far’. This step
ensures that each QFD stage is thoroughly attended to and that the current stage is
synchronised to its predecessor stage and aligned to the proceeding one.
Whilst the loop tends to slow the QFD process, it proffers the benefit of thoroughness which
ultimately makes the QFD internally compact and coherent and externally more effective and
efficient. The flow just helps in showing an abstracted logic, otherwise after Voice of
Customer and Team Establishment the stages can be followed in any sequence that suits the
organisation. It is possible to run some of the stages simultaneously. Where this is done, care
must be taken to guarantee that the system of analysis does not get cluttered with data. The
113
basis of strategic analysis is that the organisation will be virtually running a self-assessment.
The step that would then be first should be strategic planning.
Analysers. These are strategists who would take lots of time looking for details,
analysing the detail maybe at the expense of quick action.
Defenders. These are managers who look more like settlers and they basically believe
that unless something compels them to change ‘don’t fix anything until it’s broken’.
Reactors. These are strategists who are less proactive and want to cross bridges when
they come to them. Most of the time like their ‘defender’ blood cousins they would
ask ‘why buy the raincoat before it begins to rain?’
Prospectors. These are strategists who are outbound and always looking for
opportunities and markets, and chase them until resources tell them to cool down their
ambitions.
114
Figure 3.7: The 13 stages of QFD as synthesised from the literature
115
To avoid dysfunction it is important to allow extensive and mature deliberation led by a
skilful negotiator with acumen for consensus building. Organisations that quicken the
strategic planning process play by the rule of ‘eliminating problems’, but lose the resource of
multiple perspectives, commitment and the power to act on decisions. There are many
benefits that accrue from strategic planning.
Strategic planning helps the organisation to visualise the overall goals and how they
disaggregate into sectional objectives. The goal and objectives decomposition process
highlights the link between the micro and the macro (Bevington and Samson, 2012:181)
making it become more conspicuous. Strategic planning helps in environmental analysis,
selection of generic options, aligning resources allocation to objectives and the development
of control and feedback systems (Pearce and Robinson, 2009:11). Further to this it proffers a
structure within which decisions can be made. Using strategic planning tools like gap
analysis, process alignment, issue alignment and the consideration of alternative options
people gain a sense of involvement and inclusion, which Senge et al (2005) describe as
presencing.
The idea in QFD is more about edifying an infrastructure for strategic capabilities that will
sustain continuous improvement rather than chasing dead ends. Unfortunately most
institutions fail because they hire managers, give them (dead end) targets to follow on and
promise a target-based performance appraisal. This brings in a ‘humpty-dumpty’ kind of
effect on the organisation, making management rigid and inflexibly focused on their contract
goals. More often such managers muddle in selection processes as they attempt to fill the
organisation with people who support them rather than those with competences. They
actually would tend toward micromanagement, a vestige of Scientific management which
staff and academics don’t find tolerable (Oakland et al., 2002).
QFD uses tools many of which need special skills and expertise. QFD should be successful
where strategy planning is teamwork-based and focused on how best each aspect of the QFD
approach will be dealt with, including its resourcing. Bryson and Alston (2005) identify a
number of targets for the strategy planning process. However, Shah and Nair (2014:145-146)
116
studying strategic planning at Australian universities observed that there was little knowledge
and inclination to layout long-term strategies at the institutional and departmental levels. In
the QFD context, everyone is a strategist despite the fact that some persons get the title of
leader or QFD facilitator, thus becoming the chief case owners.
QFD strategists and case owners should therefore be good at traversing between and across
the roles of strategist as entrepreneur and goal setter; as analyser and competitor; as a
decision maker; as an implementer and controller and as a communicator. This is detailed in
Table 3.1 below. Senge et al. (2012: 121) writing about teams say that the quintessence of
teamwork is not about seeking mutual replication among team members, but about talking
across differences and still getting the amount and quality of strategic bundling that is
necessary to get things thought over and done effectively and efficiently to the delight of the
customer. While cohesion is necessary all the same the team must be careful of the
groupthink entrapment or their degeneration into groupthink. Team activities should
normally be defined in the team’s goals, with standards, measures and reporting
requirements. In most non-QFD environments standards, measures and reporting are but
ritualism (see Trow, 1994; McInnis, 1998). In QFD environments, these are present as datum
117
for continuous improvement purposes. What seem very necessary are QFD teams that work
right from programme, through Faculty and university to national and regional levels. With
such an infrastructure the potential to increase speed of diffusion of QFD efforts would go
up.
Table 3.1 show QFD team members in different strategic roles. Firstly as entrepreneurs and
goal setters with the role of identifying new courses of action, providing leadership, building
an institutional concept by way of developing a roadmap for quality-based competitive
advantage. Secondly as analysts and competitors who by analysis and diagnosis build robust
market intelligence from all voices in the market. The third role sees the QFD team as
decision makers who make strategic choices by way of assessing strategic options, gluing the
organisation and managing risks. The forth role call on QFD team members as implementers
and controllers who ensure successful implementations by measuring and reporting and
monitoring for effective resources utilisation. Finally is the QFD team as communicators,
ready on feedback, clarifying and communicating changes in organisational directions. The
table shows a far outcry from the traditional identity of the professoriate as in most
archetypal redbrick HEIs.
118
Table 3.1: The many hats of the strategist / manager (Scott, 2008:70)
119
QFD is essentially management by facts. To sustain cohesion, team members need a rather
strong culture of mutual respect where each feels valued by the other and communicative
patterns value the individuality and diversity of people. Senge et al. (2007:48) in The Dance
of Change highlight three new team factors that guide profound change. One of these is
openness without the fear of embarrassing questions and with the idea to developing a
genuine spirit of trust and of inquiry. The second is about localness and genuine inclusion
that encourages and creates the ‘white space’ to take decisions at the lowest possible level of
the hierarchy. Thirdly is the intrinsic motivation and proclivity to respect people’s natural
disposition to learn. According to Hoegl and Gemuenden (2001: 436), communication,
coordination, balance of member contributions, mutual support, effort and cohesion
constitute performance-relevant measures of intra-team interaction. Bureaucratic systems are
not without team-like structures such as committees, cycles, etc. Such team-like structures
differ from QFD teams in that in the case of QFD teamwork quality is subordinated to Voice
of Customer and to customer satisfaction. In bureaucratic organisations teams can easily
become tools for maintaining the status quo and championing the individual or sector rather
than the needs of the customer base.
121
3.17.4.3 Variation in meaning of Voice of the Customer
What Voice of Customer achieves in QFD relates to or is equivalent to what other quality
models like the ZimCHE, MBNQA (Malcolm Baldrige National Quality Assurance),
ISO9000:2000 (International Standards Organisation, 2008), EFQM (European Foundation
for Quality Management), IIP (Investment In People), etc refer to as customer / stakeholder /
market focus. There is a general agreement in literature that collecting Voice of the Customer
is critical to a customer-focused model and strategy. In QFD contexts what matters is not
gathering data but what is done on and about the gathered data. At the strategic level, data
collected should be about the stakeholder and not only from the stakeholders. Secondly, data
collection is an ongoing process that uses multiple instruments. Thirdly, collecting data is a
means to understanding and building knowledge rather than a goal in itself. Research work
exists that has claimed to be QFD research work, when what they refer to as QFD is the
collection of data from various populations (Chan, 2009). There is a huge difference in using
QFD as a methodology and using QFD’s Voice of the Customer. When Voice of Customer is
used as a tool or stage in QFD work, then it qualifies as the entry point (first room) into QFD.
The weakness with most educational strategies that claim to focus on the student, elevating
him to the position of primary customer, is that they collect data from the student, data that
will not help in improving relations, processes, structures and institutional customs and
practices.
In thick QFD environments the data is gathered from different key sources, processed as per
source and then globally with the aim of building webs of knowledge that will inform the
theory of the organisation and influence practice. As explained before, after many years of
research and efforts to try and be more student focused, literature is still littered with
examples of gaps between curricular and the post-qualification performance of graduates. For
the sake of thoroughness, most QFD efforts use Six Sigma tools and methods to capture
‘Voice of Customer’, as exhibited in Figure 3.10 below. Various tools can be used in
gathering such voices, including going to the workplace (gemba) and observing practice.
When such data is collected, it should be analysed and validated per source.
122
3.17.4.4 Escalating Voice of the Customer into Six Sigma Roadmaps
Bennett and Kane (2014:129) observed some weaknesses in Voice of the Customer processes
including inadequacies in data collecting instruments and in the surveyed’s lack of
intelligence, linguistic competences and skill of handling the Voice of the Customer process.
Because of these weaknesses in the collected data it should be necessary to use Affinity
Diagrams, Tree Diagrams, Conjoint Analysis and Kano’s model in order to reduce the
negative impact that the information could cause. Figure 3.8 shows a Six Sigma approach to
‘Voice of Customer’ in which the ideas, concerns of employees, the Quality Regulator,
Research and Development, students and other stakeholders are brought together for further
feed into strategic roadmaps. Affinity Diagram, Brainstorming, Tree Diagrams can all be
used to process the data. The outcome would be a series of comment types, for instance:
solution-seeking comments like ‘... hard copy modules are cumbersome’. Such comments
seek solutions to the weight and importability of the current module. It is important thus
to use other data to infer what the term ‘cumbersome’ could mean.
Figure 3.8: Six Sigma approach to Voice of Customer (Ficalora and Cohen, 2009)
123
the measure of something, for instance ‘... the module lacks practicality’. There is a lack
of a practical orientation to the level desired by the student. This may mean the student
expected more practical activities and more real-life examples.
target or desired state, for instance ‘... wish the module touched on real corporate
governance issues’. The word real may mean the content is superficial or textbook based
with no felt relevance, or may mean content is dated.
felt needs, like ‘... time is important for the module to sink’.
These examples show the value of translating the voices. Rarely would customers express
their comments or judgments in a lexicon that is of automatic value to the QFD team. It
should, however, be possible to draw immediate value from some comments at their raw
stage or after the Affinity Diagram / Brainstorming sessions and Tree Diagram stage. The
data should however be passed through Kano’s model. One of the key deliverables of Kano’s
model is the gradation of needs sifted from the Voice of Customer into: Type B (basic)
attributes or expected or must be; Type O (one dimensional) attributes; Type E (excitement)
attributes; Type I (indifferent) attributes, and Type R (reverse) attributes (Franceschini,
2002:47).
With this amount of data the QFD team can go up a further step and display the information
linking it now to key strategic issues and ‘critical to quality’ (CTQ) elements. The strategic
value of the data and the processed ‘Voice of Customer’ thus far is its usability in exhibiting
the gap between current institutional state and the targeted state with respect to each
translated Design for Six Sigma (Voice of the Customer + Voice of Business) element. Table
3.2 exemplifies this relationship. For example, from a number of ‘verbatums’ / ‘attributes’ or
voices, the QFD team infers that students want an e-module that can be uploaded on the
tablet, laptop, smartphone or iPad. This would reflect their want and need to be ‘ubiquitous
learners’. However, for the higher education institution the point of prime interest could be
profitability which means the e-module must be paid for. There would also be copyright
issues which could mean the e-module being designed such that it cannot be transferable.
The related solution is to attach the following features to the e-module: a login pin, a four-
month revival period, and an option for a maximum of two downloads per student. This
124
example shows how a QFD team can innovatively move from voice of student / customer
through Affinity Diagram, Kano’s model to an innovation that delights the customer /
student. Assuming that HEIs would take and implement Voices of Students raw and
unilaterally is rather naive. Everyone’s voice must be integrated with all others and given
meaning through the Six Sigma roadmaps.
This example looks at the Voice of Business and Voice of Customer as a synthesis rather
than raw phrases from Business or the Customer. In the example, three types of comments
are shown: solution, measure and felt need. Other comments like target can also be
extracted from Voice of the Customer. This stage in working with collected data begins the
interfacing among other QFD stages like ‘Goal Setting’ and ‘Target Values’. These links
provide the institution with a chance to view and test the link among its objectives and work.
Establishing the objective-work matrix has the benefit of precluding the introduction of work
activities that do not link to customer satisfaction. The matrix also prevents the procurement
of resources that will not be used by any sector of the HEIs in question. Academics in HEIs
have been reported as complaining about overloads, about working on ‘rituals’ that don’t
benefit anyone (Anderson, 2006). There are other issues of misuse of funds, particularly in
acquiring materials that have no good use to the organisation and its later lose through theft.
125
Table 3.2: DFSS – a Six Sigma approach to Voice of the Customer (Ficalora and Cohen, 2009)
126
The prime aim of Design for Six Sigma (DFSS), is to design and develop new products and
services or improving extant ones in ways that eliminate ‘noise’ or undue work. Figure 3.10
show the family of the Six Sigma roadmaps. Notice here that the Six Sigma roadmaps arise
from permutations of Voice of the Customer shown in Figure 3.9.
Figure 3.10: Six Sigma roadmaps linked to Voice of Customer (Franceschini, 2002)
Worked correctly each Six Sigma roadmap seeks to focalise the organisation on enhancing
customer satisfaction by continually improving quality. As a family each enhances the
contributory effect of every other towards improved customer satisfaction. These roadmaps
can be deployed within QFD to accomplish a number of goals. The advantage in using QFD
is that the team is able to link customer requirements in a very strategic manner to the
organisation’s work and resources. I explain how each roadmap operates.
127
Design for Six Sigma (DFSS) by combining Voice of the Customer + Voice of Business
The failure to combine Voice of the Customer and Voice of Business is one deficiency that is
ubiquitous in quality and quality assurance literature. Consequently students are not clear
about how to rate the gains from their teaching-learning experiences.
There is so much literature that speculates on the value that different stakeholders to higher
education can bring to the business of higher education (Altbach, 2009; Mukaddes et al.,
2010; Little and Williams, 2010). Very little is said on how these can be made to work. QFD
helps to make quality assurance agencies appreciate their work and that of institutions they
oversee. Interactions among the stakeholders help to build a systemic and systematic, holistic
understanding of the higher education landscape.
GAP(ar-d) = Qar – Qd
Where Qar is hypothesised quality and Qd is planned quality.
Most other models leave the organisation working piecemeal on the Voice of the Customer
data. QFD collates the data, turns it into management and marketing intelligence, deposits it
in the organisation’s database for future use. This is how QFD writes a concise history of the
organisation. Most quality assurance models do not tell or give advice on what should
happen with the collected data. It is not enough to inspect for a database, what matters is
what has been done further to its collection. Lamentably, for some management the
collecting of data is a goal / objective in itself. Voice of the Customer is a decision tool rather
than a fin, an outcome in its own right.
Ficalora and Cohen (2009:85) and Mukaddes et al., (2010:1) discuss some of the techniques
used in gathering this data (including surveys) and in processing it (including Affinity
Diagram and Tree Diagrams). Dealing with creating strategic capacities (Bevington and
Samson, 2012) and meta-competences (Dowling and Henderson, 2009) rather than seeking
costly and expansive structures for control would improve quality in HEIs in ways that
continuously approximate offered quality with hypothesised and expected quality. Altbach et
al., (2009:109) and Little and Williams (2010) enumerate ways in which, internationally,
students are getting involved in institutional governance as representation in departmental
committees; participation in surveys, circles; student representation bodies that have a say in
institutional governance and accountability mechanisms; membership of external institutional
audit teams; membership of other bodies outside the institutional ambits like press media,
etc. In QFD it is not enough, the important thing is the clout the students will have in shaping
policy and practice across not only student issues but the totality of education, both
downstream and upstream. The contention of the thesis is that unless the voice of the student
129
coalesces with all other voices their lot before and after their studentship will not be satisfied
by either their participation nor by the grace of other stakeholders.
Six Sigma Process Design (SSPD) by combining Voice of the Customer + Voice of
Employee
The primary aim of SSPD is to reduce the number and scope of processes in ways that
maximise value to customers. Processes are the key media through which organisational
resources are transformed in pursuit of customer requirements. It is crucial that all forms of
organisational processes from person-to-person right up to system-to-system level are aligned
and integrated in ways that, for instance in HEIs, optimise products and services value as
perceived by the student, society and industry. One approach common in QFD is running the
Six Sigma Process Design roadmap through the following DMACDOC (Define, Measure,
Analysis, Conceptualise, Design, Optimise, Control) processes. I explain each procedure
below:
Define is about defining the challenges that the current process, the key stakeholder, the
programme, the courses and other products and services are facing. It is important that each
constituent / stakeholder group define its challenges, suggest how they are caused and
propose how they can be surmounted. What is most often lost at these stages is the principle
of ‘presencing’ in the difficulties of the other stakeholder, of shared ownership of the
challenges and the solutions and that of vulnerability (that I should be blamed where I
deserve to be). In chapter 2 I refer to ‘triple focus’ as the ability to see things from my inner
self that of the other person and the common. This approach in solving difficult issues allows
for what are generally referred to as third-way strategies. Defining challenges opens up for
the SWOT analysis and organisation’s Gap Analysis and doing it correctly should take the
organisation half-way in strategic planning as it flags constraints to quality.
Measuring is about sizing the discrepancy between extant products and services
performance and the expected. Tools such as Failure Mode Evaluation Analysis (FMEA) can
be deployed to help to work through this step. The management best practice principle of
measuring and reporting delivers on this aspect. Running this step well should expose the
130
discrepancy between the offered quality (Qd) and the expected quality (Qr) which can be
expressed mathematically as below.
Q(d-r) = Qd – Qr
Where:
Q(d-r) is the discrepancy between planned quality and expected quality.
Qd is planned quality, and
Qr is expected quality.
Franceschini (2002:145) attributed this discrepancy to some seven reasons relating to
structural dysfunctionality, interfacing, mental frameworks, technology and power distances.
The discrepancy can be due to a ‘confused and approximate definition of the roles, duties,
and objectives of quality’, a phenomenon that is all too common among academics, between
the quality assurance agencies (QAAs) and academics; between academics and management;
and academics and the student.
The second is the presence in the educational supply chain of staff lacking in KASUBB
(knowledge, attitude, skills, understanding, behaviour and beliefs). A third source of such
discrepancy would be the inescapable presence of interfaces between roles, structures and
processes within and between the institution and other provider organisations. It has been
confirmed through research that the presence of obsolete technology within the various work
areas contributes to the gap between planned and offered quality. Inefficient control and
evaluation systems also contribute to creating or widening the gap. Alstete (1995), Anderson
(2006), Coates and Seifert (2011) noted that tall and bureaucratic structures create high
proclivity for such discrepancies to occur. And lastly, as observed by Hoegl and Gemuenden
(2001:436), poor teamwork quality (TWQ) encourages such discrepancy to manifest. The
ZimCHE quality assurance blueprint requires that HEIs provide evidence of stakeholder
engagement.
Analysis is about analysing the products and services to identify opportunities for their
improvement. Tools such as surveys, research and development and benchmarking can aid
this stage of Six Sigma Process Design.
131
Conceptualise is about conceptualising strategies for improving Customer Satisfaction
Performance, which involves interface mapping, strategic categorisation, and designing sets
of strategic capabilities.
Designing is about designing processes that meet the Customer Satisfaction Performance
levels. In (high) power-based institutions trouble would arise at this point. Ideally, process
design should be informed by what the QFD / design team recommends, based on the
‘Define, Measure, Analysis and Conceptualise’ done thus far. Thus decisions should be
knowledge / expert based, but in reality it is more often the case that management want
institutions run their way, with or without the data from definition of strategic issues, the
measures of their current and desired levels, analysis and conceptualisation of the scope of
the way forward. In QFD situations, process design should be subordinated to the WHAT /
needs – HOW / means matrices: What do students and other stakeholders want and how do
we deliver on those requirements? HEIs should be happy saying: ‘‘The customer said – and
we did that’’. With this kind of alignment institutions would come closer to closing the
overall quality gap: the discrepancy between expected quality (Qa) and perceived quality
(Qp).
This gap is fundamental in quality assurance as it is a global summation of the other gaps /
discrepancies ‘‘... that emerge in the planning-production-presentation-delivering processes
of a service’’ (Franceschini, 2002). If Quality Assurance mechanisms and agencies are
established to make sure or at least show the stakeholder world that they are doing something
about making perceived quality meet expected quality then they need to do more about the
sources of these quality gaps. Interestingly, the sources of these gaps lie within the controls
of the organisation and would be best addressed by the teaching institution itself. However,
an external Quality Assurance agent can enforce that conditions for the processes that should
dissolve those gaps are present in HEIs.
132
Optimise is about optimising processes and services for effectiveness, efficiency, and
flexibility with the appropriate metrics, targets, and specifications that facilitate controlling
for quality.
Control is about controlling the processes and service through a regime of premise
control; strategic surveillance; special alert control, and implementation control (Pearce and
Robinson, 2009:411-412).
Six Sigma Process Design aims to guide ‘‘... the analysis and design of work flows and
processes within and between organisations’’, say Davenport and Short (1990).
Organisations adopting the Six Sigma Process Design roadmap should gain the benefit of
being more market focused and creating space for more cooperative, collaborative and
communicative behaviours among staff. The implication of adopting Six Sigma Process
Design methods is that the work unit transforms from department to team and key
organisational figures change from being functional executives to process owners and QFD
facilitators (Silvestro and Westley, 2002). This transformation is a seal of competence in
market-oriented organisations.
Marketing for Six Sigma (MFSS) by combining Voice of Market + Voice of Business
Marketing for Six Sigma incorporates ‘Voice of Business’ and ‘Voice of Marketing’ with the
intention of making the needs of an organisation’s business requirements and the marketing
requirements succeed, particularly with regard to launching new products and services and
carving out new market segments. MFSS can operate at different levels in HEIs. Aggressive,
prospector and analyser type QFD teams can use market trends intelligence to picture up
current and predict future trends in programme, course and subject content needs. They use
this intelligence to innovate and improve curricular with a consequent strengthening of their
market and competitive positioning. It must be appreciated that business aspirations of self-
funding institutions and programmes cannot be ignored. Quality Assurance schemes that
ignore the marketing and business orientations of Business Schools will not be found friendly
by the Business school. For instance, in 2012 the CUTGBS proclaimed that it aimed for an
increase in its market share by 20% and that it aimed for profitability. It is becoming a trend
133
for Business Schools to behave as profit-seeking corporations with quite a number now
registering on stock markets. Marketing for Six Sigma uses sciences, statistics and data to
understand the needs of customers and trends regarding products and services. Once QFD
has helped institutions to achieve their business and marketing goals it is important that the
institution reinvest by increasing their resources base and creating and acquiring more
strategic capabilities.
Technology for Six Sigma (TFSS)by combining Voice of Market + Voice of Employee
Technology for Six Sigma incorporates ‘Voice of Employee’ and ‘Voice of Marketing’. In
most cases educational institutions do not have knowledge of trending technology, which
could be used to the advantage of course delivery. Technology for Six Sigma requires the
technologists’ voices on technology capability coupled with market requirements in order to
certify and release new technologies to the product designers (Ficalora and Cohen, 2009:29).
The distinctive idea in QFD is the analysis, validation and certification of technology that
will enhance the Customer Satisfaction Performance of the institution. The implication of
Technology for Six Sigma is that organisations should be able to enhance their quality
performance and gain on everything that goes with competitive products and services. One
example is the design and delivery of e-modules and in meeting the ever-increasing demands
to be ubiquitous digital learners using smartphones, tablets and iPads.
One of the problems in the translation of requirements into projects and programmes is the
variance, real or perceived, between a requirement and the values of the implementing
institution. Blackmur (2005:88) writing about the South African Council on Higher
Education (CHE) and quality in MBA programmes says that the model adopted by CHE has
a number of shortfalls, including conceptual inclarity, structural inconsistencies and that it
fails to incorporate the market in the determination of MBA quality. Regulatory agencies that
are compliance directed may cause enormous frustration on HEIs that are innovative and
want to try new schemes. On the other hand, a too-open regulatory framework may soon lose
sense of its being.
A regulator may need more the skills of balancing her enabling role and of collaboration to
build infrastructures for quality improvement. Most quality regulators emphasise and master
more the skills of monitoring and controlling than of facilitation, enabling and probing. In
chapter 2, Figure 2.1 we portray such behaviour as long old fashioned, no wonder Haggis
(2009) describe most quality assurance efforts as retrogressive. From Dale (1999), through
Franceschini (2002), Bergquist et al., (2004:248), Sallis (2002:18) to Ficalora and Cohen
(2009), the advantage of QFD is that it helps organisations shift quality assurance
responsibilities from the external regulatory architecture to the fabric of the organisation at
both the philosophical and the techno-operational levels. At both the institutional and supra-
institutional levels there tend to be issues of definitional inconsistency: What is quality and
how should quality assurance be handled? What should be the credentials of a quality
135
assurer? Lenn (1992:1), Materu (2007:3), Dill (2007:1) and Altbach et al. (2009:53) discuss
these complexities, including the need to ensure co-existence of innovation and improvement
dimensions in quality assurance. By merging regulatory requirements within Six Sigma
roadmaps, QFD pre-empts the tendency of quality regulators to focus outward on the
‘regulated’ without paying attention to their own developmental needs at the knowledge,
attitude, understanding, skills and behavioural levels. QFD can, as was observed by Aytac
and Deniz (2005:512), improve the quality of educational provision (doing the same but
better) and it can innovate (new issues / new approaches). In reference to the implementation
of Quality Assurance models in Africa, Materu (2007:24) says that the effort is constrained
by a lack of financial resources and the absence of human knowledge and skills.
The ZimCHE passes a number of requirements on HEIs in Zimbabwe. For instance, for
accreditation purposes aspiring institutions must pass the criteria of basic requirements in
four main directions: programme input; programme process criteria; programme output /
impact, and programme review.
Programme Input. This criterion covers programme design, student recruitment, admission
and selection, staffing of programme, staff complement, teaching and learning strategy,
student assessment, policies and procedures, infrastructure and resources, programme
administration services, postgraduate policies, procedures and regulations.
Programme Output and Impact. This criterion has two dimensions: student retention and
throughput rates and programme impact criterion.
136
Frameworks that regulators use for judging the quality of an institution may reflect one or
more of the criteria of quality as excellence, as fitness for purpose, as fitness of purpose, as
enhancement or improvement, or as transformation. Requirements may emanate from all
other ‘voices’.
The Planning Matrix derives from the Voice of Customer, of Marketing, of Employees, and
of Business. The broad-based approach gives the team that develops the product or service a
systematic method for comparing the performance of the current product or service in
meeting customer needs as against competitors (Chan, 2009:43). Secondly, the Planning
Matrix facilitates in developing a strategy for customer satisfaction that optimises the
institution’s ability to both sell the product and keep the customer satisfied. Thus the
Planning Matrix, in the hands of competent academics and staff, can serve as a potent
strategic instrument facilitating strategy decision-making that flows from product planning
matrix; part-subsystem development matrix; process planning matrix and process quality
control matrix. The advantage in QFD as compared to other models is that the Planning
Matrix asks the following questions for each customer need:
How important is this need to the customer (student; industry; society)?
137
How well is the institution and the programme doing in meeting this need today through
research, teaching, community services?
How well is the competition doing in meeting this need today? The competitor can range
from self-study, freely available teaching material (online, etc.), in-house company
training, conventional or distance programmes, etc.
How well do we want to do in meeting this need with the product and services being
developed?
If we meet this need well, can we use that fact to help to sell the product? It is important
to bear in mind that QFD has both moral and economic implications for HEIs.
These questions are similar to those raised by Ficalora and Cohen (2009) and apply both to
manufacturing as to the service sector inclusive of education. The Planning Matrix helps the
institution understand Customer Satisfaction Performance (CSP), which is the customer’s
perception of how well the current product or service is meeting the customer, for instance,
student needs (Ficalora and Cohen, 2009:151).
The ‘Design Scorecard’ shown below can be used during Product Planning to assess the
overall design quality against all customer requirements (Ficalora and Cohen, 2009:98). The
Design Scorecard predicts the probability of defect level and other failure modes in the
ultimate product part-by-part, process-by-process, performance area-by-performance area
and its reliability. These tools are not available with other quality assurance models. In
essence, the Design Scorecard measures how well the institution and programme are meeting
the QFD plan, including the regulatory requirements at the overall level (Ficalora and Cohen,
2009:100). Application of these tools in programme evaluation for instance of an M.Sc.
Programme should help it absorb best qualities of other best programmes.
138
Figure 3.12: The Design Scorecard and elements of Product Planning as synthesised from
the literature
The utility of the Design Scorecard rests in its ability to identify all stakeholder needs; all
regulated components, materials, etc; and internal value-creating processes. The fact that
QFD sums up all important stakeholders’ points of view and validates them as they are
incorporated into products and services design help eliminate the contest for meaning of
quality among different stakeholders. Finally, with the Design Scorecard the team should be
able to integrate focus on shared interests of stakeholders, the best resources, optimising
processes and the pertinent regulative framework.
QFD provides a method by which the development team can record the competition’s
strengths and weaknesses alongside its own (Ficalora and Cohen, 2009:153). One of the
techniques used during comparisons of own products and services is benchmarking. It is
important that HEIs train their staff in benchmarking skills. Good benchmarking encapsulates
learning from and attempting to match or outdo the best. It should involve running a strategic
gap analysis with the aim of closing the gap between expected quality (Qa) and offered
quality (Qo) and seeking new structures of thinking and action that would leverage
innovation. Despite the huge efforts taken in trying to resemble products and services of high
performance HEIs, most institutions still underperform. Sallis (2012:101) highlights some of
the mistakes in attempting to benchmark from high performance institutions; trying
benchmark as a quick fix or panacea; spying or educational espionage; trying to pace up or
copying; educational tourism; and benchmarking for a cost-reduction exercise.
Perceptual maps, radar diagrams and comparative tables are very competent tools when it
comes to running competitive satisfaction performance. All the three techniques summarise a
140
customer’s comparative perception and judgment and how the customer identifies a product
or service in comparison with the rest on the market. The amount of detail in a perceptual
map, radar diagram and comparative table can be varied by the team using them. If these
tools are used in non-QFD environments the likeliest is that the objects of comparison are
dictated or suggested by management. In QFD these should be discussed by the customer
groups. This way QFD aligns measurement to assessment and evaluation within a customer-
oriented perspective. This practice should afford the stakeholders space to express their
expectations in terms of the learning process and in terms of the attributes developed in the
graduate. These issues relate too to Section 3.9 and Section 3.17.6 because competition
cannot be divorced from the quality of the product (PPM) and what stakehplders await in the
product.
Secondly, in terms of technical responses or substitute quality characteristics (SQC). Here the
QFD design team will be concerned with knowing what is present in the processes, habits
and methods of those who provide competitor products and services. One such concern could
be, for instance, how other institutions are using video conferencing, e-books and other
channels of communication in their educational programmes. “One of the benefits of QFD is
that it creates a structure that leads developers to ask the right product-planning questions,”
say Ficalora and Cohen (2009:156).
While literature reports an increase in quality literacy, most people still deposit their faiths
with ranking and some constructs that have been unfortunately associated with quality.
Knight (2011:14-15) warns of the canard of associating international reputation; international
accreditation; global branding; presence of foreign students and international institutional
agreements with superior quality.
As explained at Section 3.17.4 there are often problematic issues regarding mechanisms of
data collection and processing for both Competitive Satisfaction Performance and ranking
purposes (Eley, 2001; Douglas, Douglas and Barnes, 2006; Yorke, 2009; Law, 2010b;
Altbach, 2009; Brandenburg and De Wit, 2011).
141
3.17.8 Technical Descriptors
Mukaddes et al. (2010:3) define ‘technical descriptors’ as all those attributes about the
programme, including teaching techniques, modules, instructional technology, facilities,
delivery modes, etc. that can be measured and benchmarked against the competition. Every
organisation faces the incessant pressure to improve and upgrade its current set of technical
descriptors: organisational culture, human competences, facilities, processes, etc. There is
also the need to acquire novel sets of technical descriptors. QFD has an inherent mode of
showing the imminent decay and datedness of competences and technical descriptors. It
further indicates the trends along which the organisation’s technical and strategic capabilities
must be developed. In QFD-based organisations Black Belt, Green Belt, Yellow Belt training
is a common strategy for building the organisation’s .strategic capacity.
Whilst all technical descriptors are important, a learner-centred curriculum must focus more
on teaching and learning approaches. Mukaddes et al. (2010:1) however lament that the
product of education is largely intangible and difficult to measure because it lies latent in the
graduate’s knowledge, attitudes, skills, understanding, behaviours and beliefs. The Kano
Model can be used to classify technical descriptors according to their importance and
deserved priority. See Kano’s model in Figure 2.5. Salmi (2011a:6, 2011b:326, 2011b:327)
discusses some nine errors committed in trying to establish world-class universities. The idea
points that distil from his discussion is that teaching and learning are the high priority
attributes and that management easily get trapped in the desire for visibility and can shift
resources away from teaching and learning into issues not directly and immediately
impacting on learning and teaching. These concerns are echoed by Knight (2011) and
Macfarlane (2012) who critique internationalisation of the institution and curricular and
professor type respectively. In student-centred, outcome-based curricular the technical
descriptor of greatest value should be the ability of the professor-lecturer to grow the
knowledge base, transform the attitudes, and help the student attain new and sharper skills of
understanding and creating new knowledge. Management can fall in the trap of seeking
qualities that are alien to the customer requirements when recruiting staff where QFD is not
deployed. Either the lecturer is what Macfarlane (2012) called knowledge producer,
142
academic citizen, boundary transgressor, or public intellectual has a huge bearing on the
curriculum content and learning experienced by those he teaches.
Secondly, objectives proffer explicit directions that allow for consistent decision making over
the long term, thus allowing the organisation to link today to tomorrow. In chapter 2 the
thesis discussed the importance of aligning the present with the future. It emphasised
integrating the deliverables from across the organisation’s structure-structure, structure-
143
function and function-function relationships and the interfaces of the organisation. Further to
this it highlighted the need for linking the micro to the macro as a strategy for attaining
consistence, excellence and enhance value for money and value for effort. Goal setting is
about change-project management and about resourcing for the future. These management
best practice perspectives are discussed in chapter 2.
Thirdly, setting goals helps us to be aware of the link between business goals and the skillsets
required to achieve those goals effectively and efficiently. Working out the WHAT -
RESPONSE matrices is a way of turning the organisation to a fact-based philosophy and
methodology of management. QFD expands the traditional conceptualisation of
accountability to include accountability to the needs and wants of the customer rather that the
thin dictates of top management. Earlier in this chapter I discussed the concept of
accountability. In non-QFD contexts accountability is to the quality assurance agency and to
the university management. QFD is a game changer, accountability is to the student, industry
and society and goal setting is subordinated to that philosophy.
Fourthly we can observe that goals cannot be set as high as possible and at the same level of
aggressiveness or ambition. This is because each organisation has limited resources. QFD is
a prioritisation tool for prioritising not the supposed goals of management but the needs and
wants of customers. This is why it is important to validate Voice of the Customer and pass it
through the Kano model. The Kano model tell us which set of customer requirements to give
what treatment as their value to both the organisation and the customer is historically
determined in space and in time. Earlier I critiqued the internal quality assurance
infrastructure, saying that in traditional models the prerogative for the definition of quality,
for quality assurance and for strategy formulation including resources allocation and
budgeting is confined with top management. QFD, being forward thinking, customer focused
and teamwork-based thinks that what matters is what the Voice of the Customer says.
Consequently the most important purpose of goal setting is mining as much data, having a
robust infrastructure of data processing and having an organisational behaviour system that
show unquestioned respect of Voice of the Customer. This is one of the reasons why
144
organisational research on model adoption needs examine management response to the
model and its results on a longitudinal basis.
Fifthly, is the question of whether to excel in all areas. Excelling in all areas may not be
feasible based on resources needs. It may not be necessary particularly where it results in
over-quality. It may unnecessarily increase costs of the products and services resulting in
violation of the ‘value for money’ principle. More often than not, institutions survive in
conditions of limited resources consequent of which choices must be made regarding where
to place special emphasis or extra resources and where not to. The point of view of limited
resources and of selecting which aspects of a programme or other services will excel and
which will not becomes a strategic imperative.
Goal setting is both inward and outward looking as well as customer focused. In the context
of QFD, Current Satisfaction Performance and Goal are combined arithmetically to produce
the Improvement Ratio. The Improvement Ratio is a multiplication factor that effectively
scales the Importance to Customer, and thus reorders the importance of the customer needs.
The most common method for determining the Improvement Ratio is to divide the Goal by
Current Satisfaction Performance:
The Improvement Ratio becomes large when there is a large difference between the Current
Satisfaction Performance and the desired Goal. In the beginning of quality improvement
effort, many of the problems affecting quality are more visible to almost everyone and are
likely to ignite conversations for action. By their nature bureaucratic systems increase
visibility of quality issues lying within lower levels of the institution. Top management
normally coerce and incentivise to make it quicker and easier to pluck them off. As quality
improvement efforts continue and all low-hanging problems get plucked, the institution still
has the challenge of searching for and plucking higher-up and more embedded problems.
Resistance to change is more often than thought embedded in top management than at lower
levels (Senge et al., 2007 / 2012).
145
Implementation efforts should focus more and more on teaching-learning as this is the only
parameter that directly links to quality of graduates. Mukaddes et al. (2010:1) say that it is
practically the teaching techniques that help individuals develop creative, critical and
perceptive ways of thinking that sharpens their ability to define, formulate and solve
problems cooperatively and individually. A number of strategic techniques can be used in
seeking improvements: business process reengineering (BPR) of ‘troubled’ system-to-
system, system-to-person, person-to-person processes; restructuring the institution;
redesigning cumbersome processes; management by constraints; benchmarking; interface
mapping, etc.
The second question is about the regulator: what is its focus, its specific requirements
regarding products and services and what are its metrics. There are cases where there is no
specific regulator, for instance, regulation on modules. Here the institution can take what the
market considers as best practice in module production as the benchmark. However the
commoner situation is where there are a group of regulators fragmented in their approaches
and standards. Then there is the question of the institution’s ability to meet each of the
requirements from the customers, the regulator and ‘best practice’.
146
Relationships matrices help institutions make the links between objectives and means of
achieving those objectives; between resources and demands; between demands and their
relative importance. In using the Matrix Diagram the team has to figure out what the strength
of the relationship between the Technical Descriptors and the customer needs and regulatory
requirements is. Each relationship can be classed as nonexistent, weak, moderate or strong.
Most other models do not get this far regarding detailed relationship between objectives
(whats) and their corresponding technical descriptors or responses. The Matrix Diagram is a
simple but powerful tool that stands at the heart of QFD, as confirmed by Ficalora and Cohen
(2009). Ficalora and Cohen (2009:100) underscore the importance of matching the measures
of quality held by stakeholders and that the system of transferring these into metrics of
quality by the institution be well developed, accurate, and understood.
The Measurement Systems Analysis technique can be used and departs from the analysis of
all physical and informational components of the whole institutional delivery system. There
may be particular reference to variables such as staff quality, novelty and capability of
technology, capacity and appropriateness of methods, quality of materials, standards,
measurements and the institutional culture. The Measurement Systems Analysis ensures that
things are not left to fall below the minimum requirements set by professional boards,
agencies, visions and other agreements. The ‘Design of Experiments’ proffers the only means
by which many relationships between products and services performances, materials,
components, subassembly specifications can be determined (Ficalora and Cohen, 2009:103).
QFD work relies on an assumed relationship between the ‘Whats’-objectives, requirements
of customers and the ‘Hows’-responses, means by which the institution accomplishes the
Whats and all other matrices whose construction depends on that relationship.
The success of QFD plans and subsequent strategies depends on the assumed product-to-
component or product-to-process relationships developed, tested and confirmed by those at
the coalfaces of each value activity (Ficalora and Cohen, 2009:103). The relationship matrix
is a new business development analysis matrix used to analyse the market to find new
business opportunities in difficult times. Institutions wanting to stay ahead of the game use
this instrument too. From an understanding of the relationship between various institutional
147
factors in shaping the quality of institutional products and services the following question
was generated: Which relationships has the CUTGBS found to be positively contributing to
quality improvement? How has staff exploited positive as well as negative relationships?
Houston (2010:178), referring to New Zealand observed that academic audit procedures had
become ritualised and therefore had lost meaningful impact on the process of assuring
quality. Problems can arise in the implementation of academic audits. Among these
difficulties are: focus; audit self-studies; conducting visits; reports; follow-ups; training of
audit teams, and selection of audit teams (Dill, 2000:xx). Audits have a huge impact on the
quality performance of the institution by helping them to evaluate how well they are fairing
against their own stated aims, goals and objectives, mission and stated standards (Hayward,
2006:50). Audits have helped institutions to vitalise their quality assurance systems and raise
issues of teaching and learning up for top management considerations. Audits have helped
institutional leaders develop institution-wide quality cultures and provide institution-wide
information on best practices. On a marketing note, audits help in proffering visible
148
confirmation that the institution is seriously paying attention to academic quality assurance
(Harvey and Williams, 2010:8).
The most common deliverable of an institution analysis is a SWOT table like the one in table
3.3 below.
Table 3.3: SWOT Analysis with resultant strategies as synthesised from the literature
This SWOT Analysis table lists under each header the corresponding points. A plethora of
actions will be taken in respect of each point on the list. The many actions may result in goal
conflict, further organisational fragmentation because each sector responses in its own ways
to its opportunities, strengths, weaknesses or challenges. Other weaknesses may involve
sectors which may be assigned responsibilities they feel don’t belong to them or they may
feel that they are not skilled or supported enough to carry out the assigned responsibilities.
The effectiveness of HEIs in developing and transforming depends on their strategic ability
to successfully engage and articulate internal and external competences. Tensions often flare
up between quality assurance agencies, governments, institutions and external decision
149
structures. Intra-institutional conflicts flare up occasionally between sectors and individuals
and if not skilfully handled they may create difficulties in quality performance. Institutional
difficulties may be resources-based where the institution does not have sufficient resources to
pursue its well-intended strategies. In some instances working with QFD may require the
construction of long tables and personnel may not have the skills to handle and analyse such
tables. Working with many objectives and across numerous people and processes is a task
that needs sharp negotiation skills, energy and patience. Good negotiation skills are not as
quick coming and common as is the temptation to ramrod one’s way through.
There may as well be misfits between institutional culture and strategy and between
institutional strategy and structure. Running through the process of drawing up the
Relationship Matrix the team improves its awareness of the institution’s strengths and
weaknesses. A SWOT Analysis can be run at this point or the team may decide to draw an
order list of the difficulties as relating to external threats and internal weaknesses. If the team
does not have guaranteed clout to influencing other resources at the system and strategic
levels it can go as far as listing the institutional difficulties and making blow-by-blow
recommendations. Institutional assessment must help the institution identify and frame
strategic issues. This can be attained through identifying individual strategic issues, creating a
master list of key strategic issues and drawing an institution-wide strategic issue statement.
With a complete understanding of the strategic issues it becomes of paramount importance
deciphering strategic issues from the operational issues (Bryson and Alston, 2005:85). To
achieve maximum gains from QFD, management must drum the importance of quality in the
institution’s strategic plans.
There are various sources of data that the QFD team can peruse to get a thicker understanding
of competitor products and services: one source are recorded data that includes the
organisation’s annual reports; media articles; press releases; regulatory reports; special
presentations; analysts’ reports, and speeches. The other source is observations of patents and
alumni performances. The third source is the opportunistic data sources that include
meetings; seminars; trade shows; workshops and recruiting ex-employees of the target
competitor.
In QFD, Target Values are themselves targets for continuous improvement and are addressed
by the Six Sigma Process Design roadmap. In the context of QFD, the determination of a
Target Value is a result of the interaction among customer needs, the current institutional
performance and the competition’s performance. In this context a Target Value become the
object of benchmarking. When the QFD team has determined key Substitute Quality
Characteristics for each technical descriptor, the QFD team should establish the procedure,
process or activity that will deliver the particular value. A QFD approach helps the institution
to complete the value design, management and delivery processes in ways that maximise
value to customers. Once target values are established, the QFD team prioritise them
151
according to their overall importance. Kano’s model can be used to classify them into
dissatisfiers, should-haves and delighters. Student evaluation of teaching effectiveness
(SETE) can be used to help this process. The primary inputs in the setting of target values of
Substitute Quality Characteristics according to Ficalora and Cohen (2009:219) are:
drawing a rank order of Substitute Quality Characteristics into a priority list – priorities;
drawing up the competition’s technical performance – competitive benchmarks; and
the development team’s product technical performance – own performance.
After running an assessment of the competitive market, the QFD team can make decisions
for each key Substitute Quality Characteristics to:
aim to beat the competition;
concede technical leadership to the competition; and
match the competition (Ficalora and Cohen, 2009:219).
There are two basic approaches in setting target values. Mathematical modelling can be
applied where the process allows for the use of mathematical modelling. When the
mathematical modelling approach cannot be used, a different strategy should be used.
Among the alternatives are the following. The design team can treat each Substitute Quality
Characteristic as if it were the single determinant of Customer Satisfaction Performance on
an attribute. This strategy or approach of pursuing excellence is not uncommon in both the
manufacturing and the services sectors. A QFD-inspired professor would say ‘’if my students
should love this subject it should be because of its logic’’ and optimise the logical aspect of
the subject then again ‘’if student should love this subject it should be because of its
relevancy’’ then optimise the relevance aspects of the subject. She should go on until all the
determinants of curriculum quality are covered and optimised. Optimisation is itself not
quality but a roadmap to quality creation. Each subject of optimisation thus becomes a Target
Value.
The second approach is choosing the most aggressive of the target values (Ficalora and
Cohen, 2009). This approach may involve extra-organisational research work whereby QFD
team members use competitor metrics and benchmarks to pin-point what are the current
152
levels of desired performance in a certain determinant of quality. The organisation then
strives to come nearer, perform equally or even surpass the best performer on that
determinant of quality. If a market of students consider for instance lecturer qualification as a
key attractor of value, universities may soon have well paid professors and doctors
irrespective of a fragmented curriculum that has no other value than employing the lecturers.
‘Target value’ refers to ‘how much’ for the technical value and this acts as the reference
point in related comparisons. Benchmark technique can be used in improving target values.
Target values are key in determining the direction and magnitude of Improvement Ratio. In
chapter 2 I discussed the management best practice of measuring and reporting and it is
important that the current situation be assessed, quantified and collectively people measure
how far they want to stretch. Only through stretch goals can target values be achieved.
153
Figure 3.13: Correlation matrix building into the House of Quality (Akao et al, 1987)
Institutions using any of the quality models available on the higher education market must
address how the programmes’ quality objectives are related to and are translated into
operational plans of the various work units. Aligning the goal breakdown structure (GBS) to
the institution breakdown structure (IBS) and the work breakdown structure (WBS) is an
enormous strategic accomplishment. The team need to look at how the quality values are
integrated into the management system and what the programme’s approach is for each type
of competitive comparison.
154
the attributes or parameters that matter most in the eyes of the stakeholders, students and
industry. The Critical Parameter Management is used in managing the relationship between
products, services, subsystems, components, materials and process parameters that derive
from designing market-oriented services and products. The criticality of an attribute can be
determined by its newness, importance, or level of difficulty, as determined by Ficalora and
Cohen (2009:98). The Failure Mode and Effects Analysis – Design Emphasis is used to
identify and thus prevent defects and disfunctionality - a hallmark of the Theory of
Constraint.
Every failure mode produces some kind of negative effect on observers and users of products
and services. It is therefore important that each worker understands the criticality of each
failure mode on the customers and stakeholders. Criticality is based on three attributes. The
first attribute is the probability of each failure mode occurring and this is denoted by the letter
P. The second attribute is the seriousness / criticality of the failure which is denoted by the
letter S. The third attribute is the difficulty of detecting the failure before the product or
service is used by the customer and is denoted by the letter D. Whilst most people would
think that this stage of QFD is too technical and has no space in the services sector the
opposite is true. In higher education where the professoriate design the course outline, does
the teaching and marks the examination lots of failure modes can remain embedded latently
in commissioned process and may take quite long before an unusual incident spills the bean.
In chapter 2 I talked of the management best practice principle of ‘being up front’ and
emphasised that business and ethics should be inseparable. A content misconception may run
for years undetected as is inflation of marks for some individuals.
Each of P, S, D is rated on a scale from 1 low to 10 high. The criticality index (or risk
priority number) for each failure mode can be calculated using the expression:
C = P x S x D.
The numerical value of the calculated C has a strategy overtone on the institution. For
instance, a C = 10 x 10 x 10 = 1000 will have a different impact to a value C = 1 x 2 x 1 = 2.
The value to customers of various product or service attributes varies in space and time.
155
Unlike the Kano model, more often the classification of customer expectations through the
‘Absolute Importance’ route yields three categories. One of them is the ‘Taken-for-granted or
fitness-to-standard. With such attributes improvement beyond a banal level is not necessarily
followed by further customer satisfaction. This is most common with most of the regulatory
requirements. One way of delighting customers on these expectations is to innovate around
them rather than improve them.
The second class refers to the competitive category in which an improvement on attributes
yields further customer satisfaction. This aligns with the thesis discussion on Goal Setting
and Improvement Ratio above. In the Kano model these expectations would fall within Type
E needs and wants.
The third bundle are the delighters / surprisers or fitness to latent expectation. These are
attributes that when incorporated in a product or service enormously delights the customer
because in his eyes they have a huge ‘wow’ effect. While the thesis appreciate the focus on
customer this is not a kind of the organisation blindly following every customer demand.
This point is emphasised in the QFD stage ‘Institutional or organisational Difficulties’ where
Voice of the Customer is related to organisational capability and competition. In working out
the Absolute Importance the team must use their intuition to and run their data through the
processes of: listening for thick meaning; interpreting Voice of the Customer for its
implications; responding through alternative offering or directly and monitoring products and
services, customer response and organisational capabilities.
158
Apart from internal incompatibility, models may exhibit dysfunctional external
incompatibility. External incompatibility is, for example, when a model used for institutional
management does not accommodate principles of a model used for products and services
creation, management and delivery. One such example would be the use of QFD in the
academic areas, for example curriculum designing whilst using Business Process Re-
engineering (BPR) in the same unit for administrative issues. To maintain links among the
processes it would make sense to use Six Sigma Process Design (SSPD), which is part of
QFD. The implication of this reasoning is that when organisations choose a model or strategy
they must examine to what extent the candidate-model will cater for the structures, processes
and culture of the adopting organisation. Most evaluation tools regard quality and evaluation
as functioning separately (Mizikaci, 2006). Thus adopting QFD as a change-project model
allows a systems level evaluation of the programme.
QFD is one model that fits in with current trends in higher education, such as increased and
genuine stakeholder management rather than stakeholder manipulation (Harvey and Green,
1993; Harvey, 1996; Tierney, 1998; Tran et al., 2011; Lagrosen et al., 2004; Becket and
Brooks, 2006; 2008); commoditisation; transparency and blending. Whilst other models
enforce radical or leapfrog kinds of change the absolute versatility and incremental approach
to change taken by QFD gives implementers chances to self-assess on progress and to blend
QFD into their present infrastructures. QFD works on organisations at any scope that the
organisation wishes. This gives the organisation the opportunity for differential
implementation according to organisational priorities and resources at hand. QFD however
shares some characteristics with the rest of the models. Business Process Reengineering, Six
Sigma, TQM, Theory of Constraint and QFD seek to improve products and services through
a structured approach to performance improvement that centres on systematic design and
management of an institution’s business processes (Chang, 2006:31). A common family of
principles guide these models:
business processes are assets that are essential for creating value for the customer.
by measuring, monitoring, controlling and analysing business processes an institution
can deliver consistent value to customers and has the basis for process improvement.
business processes must be never endingly improved.
159
information technology is an essential enabler for Business Process Management.
Quality gurus like Crosby (1979), Juran (1993), Deming (1986), Franceschini (2002), have
identified what Chang (2006:31) crystallised into eight practices that husband QFD and
Business Process Management:
transform organisational structure toward process-orientation or boundary-less,
ambidextrous organisations.
replace the all-powerful-executive manager mentality with the idea of competent and
enthused process / case owners who are ready to champion QFD.
provide support at the strategic, management and technical level whilst genuinely
adopting a bottom-up perspective.
institute information technology systems to monitor, control, analyse, and improve
processes.
adopt a mentality to continuously multiskill the workforce and continuously improve
business processes.
align employee total reward system to business process performance.
utilise permutations of incremental and more radical approaches to implement process
improvements.
3.19 Conclusion
In this chapter we established the relation between quality and innovation and suggested that
this forms the DNA of quality improvement. We have further argued that because QFD can
be used as a measurement, assessment and evaluation tool it is the best model for programme
quality assurance. QFD’s value in quality assurance is further strengthened by its self-
evaluation nature and that it is holistic in the sense that each of its stages subsumes one or so
traditional and classic quality assessment and evaluation models / approaches. Ten such
exemplifications are given. We have linked the benefits an adopter of QFD would gain from
his understanding of the model and depth of its adoption and the intensity of use of the QFD
tools. This has always distinguished winners from losers with QFD. We explicated the
concept of disconnect as the absence of a link that should just naturally be there and suggest
that their presence exhibits the poor application of the Theory of Constraint. Two such
160
disconnects exemplified are that between purpose and metrics / standards, and the mentality
of accountability to those with power as a measure of quality. A point is argued that in QFD
accountability is to the customer, as it is the Voice of the Customer that defines what quality
is and suggests how it can be gotten. With this argument we shift the prerogative to define
quality from uninformed guts of management to the processed Voice of the Customer that is
escalated through the four Six Sigma roadmaps to transformation and fitness for purpose. We
have discussed the traditionally accepted value of licensing, accreditation, ranking,
professional associations and external examination as an infrastructure for external quality
control. We argued that their persistence is welcome but each’s implications for quality is in
what it does on the ground than simply by its existence. We suggest with great care that these
activities have meaning only in the framework of Voice of the Customer. By so doing we
reduce the disconnects and fragmentation with the structures-structure, structure-function and
function-function relationships in quality assurance work. In this perspective we shift the
locus of quality assurance work to the heart of the organisation’s philosophy and
methodology of doing work. This is the quintessence of QFD whose structure is detailed and
creatively embedded with quality assurance literature in the last segment of this chapter. The
proceeding chapter discusses the research methodology.
161
CHAPTER 4 – RESEARCH APPROACH, RESEARCH METHODOLOGY AND DATA
COLLECTION STRATEGIES APPLIED DURING THE INVESTIGATION
4.1 Introduction
I took a structured approach in the investigation of how staff within the Chinhoyi University
of Technology Graduate Business School (CUTGBS) responded to a ‘quality function
deployment’ (QFD) -based model for quality assurance in the Master of Science (M.Sc.) in
Strategic Management Degree Programme. This chapter shows the research paradigm,
research epistemology, research methodology and the methods interlaced within the case
study. The chapter closes with a consideration of ethical norms followed during the inquiry
and processing of data.
162
This study views qualitative research as a family of potentially interweaving and open
approaches that reflexive researchers use (taking the perspective of the research subject and
participant in their contexts to understand the quotidian orientation of phenomena), often
departing from a case, to construct elements of theory in ways that help us configure reality.
A qualitative research approach suits this study because the research questions focus on how
processes, events and structures are interrelated and co-adapt within the figuration of the
M.Sc. Programme and the ‘gang aft agley’ of operational realities confronted by the
CUTGBS (Bryman, 1989; Hogan, Dolan and Donnelly, 2009:12). I have, however,
encountered instances where I have asked participants to ascribe a numerical value to
qualitative ‘objects’ by use of something like a Licket Scale. Bernard and Ryan (2010:4)
describe this activity as the quantitative analysis of qualitative data, and they refer to the
following three variants in a four-quadrant model:
qualitative analysis of qualitative data using resources such as hermeneutics, grounded
theory, etc.;
qualitative analysis of quantitative data when we search for and present meaning in what
emerges from quantitative processing; and
quantitative analysis of quantitative data by application of statistical and mathematical
programmes.
A quantitative research methodology would fall short of revealing the precise mechanisms
and processes in a research on the adoption of a quality model (Gubrium and Holstein,
1997:11). It would not be able to illuminate the complexities in the structure-structure,
structure-function and function-function relationships in an educational institution. Literature
generally characterises education as opaque. Qualitative research, as observed by Hogan,
Dolan and Donnelly (2009:11), can produce compelling knowledge of how and why people
behave as they do in institutional, family or individual roles. Furthermore, this study strived
to understand what exactly led to particular decisions or choices strategists and management
took and technical staff made as part of the implementation of QFD. It further looks at the
triad (or triangle) of mutual influence between resources – management – strategic
opportunities. A quantitative research approach would struggle to give a convincing review
of such a complex web of interdependences (Deshpande and Webster, 1989). Historically,
163
qualitative research has been conducted by various means that encompass direct observation
of a sample, case studies, the examination of relevant texts, interviews, focus groups, e-mails,
text messages, instant messages, Twitter, and online chat, among others (Etherington,
2004:80; Kuckartz, 2014:2). The use of a majority of these techniques has put me in intimate
vinculum with the processes in the CUTGBS, participants and literature, thus affording the
research process an opportunity for data triangulation (Jackson and Mazzei, 2009:223). I
have taken part in several conversations with participants on social networks like Viber,
Skype and WhatsApp and found such conversations to be more informal, yet richly
informative. Jones, Torres and Arminio (2006:12) and Fischer (2005:411) say that qualitative
research is pragmatic, interpretive and grounded in the lived experiences of people who
interact in webs of interdependent relationships at the micro-micro, micro-macro and macro-
macro levels.
164
would be more suiting to this study than a case study approach. The constructivist movement
assigns two main roles to the researcher: an emancipator role and an interpreter role (Jones,
Torres and Arminio 2006:24). The former is not prominent in the instance of this study as its
focus is more on ‘response to a model’ than ‘response to use and distribution of powers’.
Because numbers cannot represent the experiences of the subjects of the research, the
researcher opted for an interpretive and constructivist approach in the understanding that
meaningful reality and knowledge are co-constructed from human interaction within
essentially social contexts (Hultgren, 1989:41; Crotty, 1998:42; Torres, 2004:459), in which
humans have views and act according to their individual perspectives (Van Manen, 1990:9
and Davis, 2002:511) unless compelled to act otherwise. The researcher experienced many
instances of shuttling between sampling, data collection and analysis of the collected data,
wherein each activity informed all others. The lengthy immersion in the research context
helped the researcher to make comparisons between views, situations, actions, accounts, and
the experiences of respondents. Comparisons also focused on how some respondents
‘changed’ over time. Contrasts were also studied incident by incident, between data and
category, and among categories.
165
4.4.2 The ‘theory’ of a ‘case’
Jones (2014:12) defines an organisation as ‘‘a relatively enduring group of people with some
degree of coordination around a common principle that has a more or less identifiable
boundary’’. In phenomenological qualitative research, the whole organisation or part thereof,
can be of researchable interest (Hogan, Dolan and Donnelly, 2009:35). The researcher may
choose to draw the boundary along legal, physical, relations of mutual recognition,
coordination, common principles, boundary and relative persistence (Jones, 2014:13). The
intensity with which organisations adopt certain principles, strategies, models and structures
has been of research interest across the disciplines. Numerous researches have focused on the
adoption of various management and quality principles. The reports on the success or failure
of the principles and the adopting organisations are ambivalent. Apart from focusing on the
successes or lack of strategies or models, etc., studies need to cover and draw into the fold
issues regarding for example how people within the organisations respond to models and
their accoutrements. Organisations don’t implement strategies, principles or models, the
organisation’s staff do.
Jones, Torres and Arminio (2006:53), whilst warning against the use of case study as a unit
of analysis, are of the opinion that the case study methodology is frequently used in higher
education because many aspects of the higher education system represent cases. However,
Stake (2000:435) and Hartley (2004:323), among others, suggest that a case study is merely a
choice of what is to be studied and not a methodology in itself. Yet Merriam (1998:12) refers
to the case study as an intensive, holistic description and analysis of a single bounded system;
something bordering on methodology rather than just a choice. The choice of the CUTGBS
as the unit of research analysis fits Merriam’s (1998) characterisation of a ‘case’ as a
bounded system because: it is a case of something that has discernible boundaries that
separate it from other systems in its context; the qualitative approach explicitly attempts to
preserve the wholeness and unity of the CUTGBS; data and methodological triangulation
were sustained. Care was taken not to rescind nor repudiate the many complex relationships
between the case and the wider environment from which it is ‘cut’.
166
This study uses a reflexive research methodology, which in its essence repudiates a
simplistic, lineal relationship between measurements, observations, statements of interviewee
responses, the study of archived data and statistical presentations and the contexts from
which they arise (Alvesson and Skoldberg, 2000:9). Jones and Abes (2003:473), whilst
acknowledging the peculiarities that drove the Merriam and Stake classifications of
qualitative case study, are of the opinion that overlaps can and do occur among them. The
study has a more intrinsic-heuristic value. This is because it sought to investigate ‘how’ and
‘why’ things happened within their contexts, thus on what ‘worked’ within. According to
Merriam (1998) and Stake (2000:437), intrinsic-heuristic case studies should deepen
understandings and fecund insights gained from within in ways that lead to new meaning and
a rethinking of the processes of the ‘case’ and related phenomena. Systems in triage and
institutions that are looking for breakthroughs need less theory and more of what works.
Theories tend to be more domesticating rather than encouraging to ‘border-jumping’.
Postmodern, fundamentally quantitative, research will not help in learning what works, at
least not as much as would qualitative inquiry (Gubrium and Holstein, 1997:11 and Jones,
Torres, Arminio, 2006:225).
167
I adopted a phenomenological approach to the study to explore the nature, scope and
dynamics of the process of QFD and programme quality assurance, while bracketing,
disarming and suspending my own researcher presuppositions. Researcher presuppositions
may be biased due to objective science and authoritative sources encountered in literature
review. The other source of bias arises from the imposition of criteria of validity that is alien
to the phenomenon under study (Van Manen, 1990:47). To guard against these two key
presuppositions I completed a thorough literature review and reviewed drafts of other persons
with related experience and work embedment. The abiding concern throughout the research
process was to garner as much information and data through thick descriptions of quotidian
experiences of respondents.
The relationship modelling, types of questions and questioning techniques were designed to
transcend respondent experiences and reveal the reasons behind the choice of their
behavioural strategies and what they factored most in constructing a response model to the
implementation of programme quality assurance through a QFD model. The desire to be
representative yet thorough brought the research into a situation of batting average: broader
sample and ‘thin’ data or small sample and in-depth descriptions. The focus of this study,
characteristic of phenomenological research, was to generate thick descriptions of
experienced lives. The smaller the sample, the better it is to build relationships and attain in-
depth immersion in the phenomenon under inquiry (Jones, Torres, and Arminio, 2006:49).
Van Manen (1990:77) points out that the quintessence of a phenomenological approach
resides in the procedural relationship that the researcher edifies with data sources, with the
idea of reflectively apportioning, clarifying and exhibiting the chemistry of the lived
experiences.
During the data collection process, I consciously tried to merge diachronic and synchronic
analyses with the aim of thickening data and acquiring a more figurative perspective on the
study. This also involved the use of many research questions that could be answered through
multiple data collection methods. This strategy followed on the advice of Ponterotto
(2006:538) that meanings emerging from profound analysis improve the theoretical validity
and practical implications of findings.
Figure 4.2 below indicates the strategy used to achieve the thick descriptions. Triangulation
was used around each research sub-question to afford the researcher an opportunity for in-
depth understanding of aspects of each question and thus of the whole study. See Table 4.1,
169
page 157 as it highlights the research design showing how each question was focused from
different perspectives.
Figure 4.2: Approach to thick description and validity of the research (Ponterotto 2006)
170
Despite the fact that copious amounts of data on their own do not guarantee any thicker or
better understanding of the phenomenon if not well articulated internally as well as
contextually, many authors think it is an indispensable condition for rigour in research.
Among them are Fielding and Schreier (2001:2), Patton (2002:247), Fredriksson (2004:72),
and Lillis (2012:62).
4.5.1 Observations
An observation schedule was drawn up using information gained in literature review,
document analyses and interviews. Objects of the observations were facilities at the: library,
computer laboratory, dining hall, lecture halls, and lecture rooms.
The idea in the observations was to establish the state of the facilities in terms of fitness for
purpose and quality. The ZimCHE singles out the same features when inspecting institutions
and assessing them for licensing and accreditation. In drawing the observation schedule, I
worked through the ZimCHE forms for ‘institutional accreditation’ Section 3 on
infrastructure. A checklist was used and notes were added.
Elias (2000) suggested that people respond to stimuli on the basis of the meaning they attach
to the stimuli. To avoid bias, I would seek discussions around each of my observation
session. This way I found myself getting more and more documents and people being drawn
into my sample through snowball sampling. Conversations with every new participant helped
to unearth nuances to perspectives that had not been captured in previous interviews or had
been misconstrued.
171
Table 4.1: The research design
172
Alvesson and Skolberg (2009:288) observed that even a single interview combined with
other qualified material can make a huge rendition. Interviewing is ‘‘the most familiar
strategy for collecting qualitative data’’ says DiCicco-Bloom et al. (2006:314). It is widely
used across disciplines (Vellnagel, 2010:2) and aids in discovering information about the
past, thus helping to understand the context and motives a lot easier (Fontana and Frey,
2000:657; Wengraf, 2001:3; Vellnagel, 2010:5). In carrying out interviews, I assumed that
“each participant is the single person to provide all the data on the case”. I therefore listened
to every nuance and reflexively drew each participant into telling more and more of their
many episodes. As a result, I conducted many interviews, revisited participants and re-
interviewed them.
In-depth interviews were held with participants who were at the Chinhoyi University of
Technology and the CUTGBS and have roles at strategy formulation and implementation
coalfaces. Interviews were also held with participants outside of the CUTGBS who had a
hand in the running of the CUTGBS, either directly or via the university management
system. For the sake of comparison, a parallel study was conducted with an almost identical
sample from two other business schools. The notable difference was that one of the two
enjoyed greater autonomy from the mother university, yet the other was equally struggling
with the wishful thought of being less beholden to the mother university. The questions were
drawn from the four strands of the Strategy Focus Framework and the steps of the QFD-
model:
Planning for the strategic adoption of QFD and programme quality assurance
infrastructure and responses of staff to the strategic planning processes
Making the strategies work and the responses of staff to the management’s canvassing,
inveigling and coercion.
Managing the implementation of QFD and programme quality assurance and the
responses of staff thereto
The strategic management of risks that can frustrate intentions to institutionalise quality
and the responses of staff to all such efforts
173
Efforts were made to have the participants internalise the non-political nature of the study so
that their participation would not jeopardise their employment. It was also necessary to
sensitise potential respondents so that they could be in a more informed position and be able
to give lots of data. The idea of running ‘the scoping’ is anchored in epistemologies that view
truth as materially based and as having the potential of being misrepresented in people’s
words, thus the researcher needs to be more anchored in the dynamics of the case under
study. In some instances I would share a story tactically so as to encourage participants to
disclose more or as a way to encourage deeper reflection (Ellis and Berger, 2003:162).
The venues for face-to-face interviews were negotiated with each respondent and efforts
were made to achieve confidentiality, comfort, quietness and ample time. Glesne and Peshkin
(1992), Smith, Flowers, and Larkin (2009) consider these to be effective strategies in dealing
with respondents. It came to my attention that confidentiality was not a big concern to around
half of my academic participants. They were asking for documents and telling secretaries that
they needed them for this research. Some appointments were arranged by those who had
given me the referrals.
At the end of each interview session, participants were requested to summarise key points.
The researcher would reassure the respondents of the confidentiality of the information, sum
up their 'take home' points and provide a card with a ‘thank you message’ and contact details.
174
Where mobile and e-mail contacts were availed, the researcher would send a thank you
message on the same evening, indicating particularly interesting points and giving an
indication that further encounters would be most welcome. The strategies adopted during in-
depth interviews of both focus-group and one-on-one nature were to attain thick descriptions
from respondents and improve theoretical and business implications of the research.
Apart from being cost effective (Merriam, 1988:31) document analysis is a systematic,
analytic and critical examination of tangible information used with the idea of gaining
profound understanding of trends and patterns that emerge from the recorded information
(Creswell and Clark, 2007:114; 2011; IAR, 2010:2). Staffs were interviewed against
stipulations in some of the documents. Academics were interviewed on how the standards
influenced their curricular activities. Other documents that were analysed included:
the ZimCHE ACT of 2001 to establish the mandate, powers, structure, functions and
other issues of legal values of the council;
the Chinhoyi University of Technology ACT to establish the mandate, powers, structure,
functions and other issues of legal values of the council;
the university ordinance to establish the structures, policies and targets of the sectors of
the university;
the Strategic Plan of the Chinhoyi University of Technology to establish the priorities,
future plans and long- and short-term goals and how they were intended to be reached;
the Strategic Plan of the Chinhoyi University of Technology Business School to establish
the place of the M.Sc. Programme and how project-change management would be rolled
out;
the Chinhoyi University of Technology website;
the ZimCHE website;
the National Manpower Advisory Council (NAMACO) website to establish how the
council established the national development priorities, what they were in each sector and
how they were utilised by higher education institutions;
the lecturer profile to establish the qualifications, work-practical experiences, and
publications lists of the lecturers;
176
the ‘Student Evaluation of Teaching Effectiveness’ (SETE) form used in the CUTBS; and
examiner comments and reports.
Document analysis entails would highlight areas of interest to the writer as well as to the
party to whom it is directed. The researcher approached document analysis with the aim of
finding how they directed activities aimed at quality assurance and improvement. Secondly,
this effort highlighted the variance between what is declared, intended and what is actually
done. Each document was referenced many times over in a web-like format. This was done
in reaction to ‘leads’ and clues gained during interviews. Original documents were
photocopied and sections of interest were highlighted. These are sections that represented
significant themes in relation to quality strategy, making the quality strategy work, processes
and risk management.
A two-pronged process of data analysis and theme development emerged during the data
analysis process. This process of generating process-oriented codes should help to conserve
what could be lost with a pre-conceived coding that is rigidly dependent on the research
question. I indicate this point in figure 4.3 below which I explain now. The various research
questions could best be covered in different documents. Notwithstanding, some document
would be more expansive or detailed than all others. Still all the information would be
mobilised. Apart from the document analysis relevant details in each analysis could be mined
from other sources like interviews, literature review, and peer reviews, etc. Plugging one data
with all others thickened my understanding of participants’ presentations at both the global
level as of each research question. .
177
Figure 4.3: Alignment between the research question, data sources and evidence generated as synthesised from the literature
178
4.6 Sampling
In 2012, a request to undertake this research was discussed with senior officials in the
Ministry of Higher Education and Training, ZimCHE, Chinhoyi University of Technology
and the CUTGBS. Responses were immediate and many key informants were immediately
identified together with their telephone, e-mail, and workplace details. Reaching people and
sources of relevant data is an important guarantee for completeness of the study and its
subsequent credibility. Marshall and Rossman (1999), Strauss and Corbin (1998:201)
acknowledge the benefits of snowballing as a strategy of choice where the study seeks to
make comparisons among the voices of participants, go to as many places, study as many
documents, examine many events, all with the aim of exposing variations and densifying
categories in terms of their scope and dimensions (Morse and Richards, 2002:68).
The following was the sample from which data presented in this study was obtained. I chose
the CUTGBS because I was relatively familiar with the context of the university, the
CUTGBS, and the QFD-model, which the university used to conduct quality assurance.
Glesne and Peshkin (1992:27), Jones, Torres and Arminio (2006:65), and Patton (2002:46)
affirm that the researcher’s understanding of the ‘case’ improves the sampling strategy in
ways that guarantee in-depth understanding of the case under inquiry. Gradually I found
more references and more documents, as well as many invitations from staff within and
outside the CUTGBS. In this study the researcher downplayed cost and convenience issues in
favour of a more rigorous, purposeful and strategic approach.
In drawing the sample, I was unbiased regarding age, gender, qualifications, length of
employment, or whether one was a part-timer or employed permanently. Patton (2002:244)
once said that there are no rules for sample size in qualitative research and that the sample
size is determined by what the researcher wants to know, the purpose of the inquiry, the
abiding interest, what will be useful, what will have credibility and what is doable in the
timeframe and with the available resources. In contrast to quantitative research, samples in
qualitative research are not defined upfront, what Emmel (2013) call the theoretical sampling
approach.
179
Figure 4.4: Relationship between number of interviewees and cumulative percentage of valuable data as observed from data
analysis
180
The increased communication resulted in the research reaching a point of saturation (Lincoln
and Guba, 1985:202) where I began to notice nothing new in incoming data (seen, heard or
read) relating to the categories of analysis (Jones, Torres and Arminio, 2006:71 and Meloy,
2008:154-155). Figure 4.4 depicts this. The saturation point was nearly reached with the
tenth respondent of the in-depth interviews. With the idea of maximising quality data (Jones,
Torres and Arminio, 2006:72), the following sample evolved during the research. During
interviews participants were promised anonymity and in that spirit all efforts have been taken
to guarantee such. The following codes will therefore be used:
SM1: Senior management personnel-1. SM2: Senior management personnel-2.
SM3: Senior management personnel-3. SM4: Senior management personnel-4.
SM5: Senior management personnel-5. SM6: Senior management personnel-6.
CUTGBS Academics who participated in the study were coded as follows:
AC1: Lecturer – 1 AC2: Lecturer – 2
AC3: Lecturer – 3 AC4: Lecturer – 4
AC5: Lecturer – 5 AC6: Lecturer – 6
ZimCHE participants were coded as:
DR1: Director-1 DR2: Director-2
DR3: Director-3
Student participants (individual)
STD-1 STD-2 STD-3 STD-4 STD-5
Student participants (in focus groups)
FG1-STD 1
FG2-STD 1
Alumni participants: AL-1 AL-2 AL-3 AL-4
At the same time it acted like a multiple adaptor plugging in the different data sources. I
began to think of and accept doctoral research as what I would call ‘the dance of a qualitative
research’: up-down: forth-back: left-right: in-out and a circle. Follow-ons and probes added
layers over and above data from initial interview. Descriptions became thickened over the
data collection and research period. In transcribing, I began each sentence on a new line. The
lines were double spaced. This gave me lots of blank space to jot in notes each time I perused
the transcripts. I did all transcriptions myself. I shared some of my transcriptions with a
former colleague, who is a PhD holder, a director of a business school and who supervises
some master’s degree students. My supervisors had additional points to highlight on my
transcription.
182
Figure 4.5: Movements between a transcribed interview and other data sources as synthesised
from my practice
183
frightening process – frightening in the sense that in many of the instances it felt that I had to
pronounce my position, my own ‘micro theory’.
Mason (2002:150) points out that coding procedures that are built on pre-constructed codes
are convenient but that they haemorrhage otherwise important facts that are prima face not
relatable to the codes. The other weakness of this approach is that authors of the documents
and those who answer research questions are elevated to levels of experts and of authorities.
Kvale (1996, 2007) points out that the data analysis process ultimately brings the researcher
to a decision point of many alternatives. Each alternative would have some bearing on the
global validity of the presented data. Apropos the issue of validity and transcription
Etherington (2000:292) highlights the advantage the researcher would reap out of doing the
transcription personally.
The nature of the principal research ‘question’ required the research to focus on the QFD
model and thus the data collection questions and data analysis had to pay unfretted attention
to the steps of the QFD model: strategy planning and structure; establishment of the cross-
functional team; voice of the customer; regulatory requirements; product / service planning
matrix; Customer Satisfaction Performance; Competitive Satisfaction Performance;
Technical Descriptors; Goals and Improvement Ratios; Relationship Matrix; Institutional
Assessment; Competitor Product / Service analysis; Target Values for Technical Descriptors;
Correlation Matrices; and Absolute Importance.
Analysis of the intensity with which the CUTGBS treated each of the above steps constituted
much of the follow-up interviews and in this case questions were more direct, even though in
a few cases answers were elusive. In the conversations, as in the coding, I did not concern
myself with the exact words I have used in the framework or the logical order of the above
13 codes. Alternative wording or even descriptions were very welcome and were read for the
appropriate / commensurate code and stage. Otherwise a more open coding approach was
used in the large corpus of the data.
184
The advantage this research sought in taking a process-dependent coding is that it is more
representative of reality as a mobile network of interrelated and interdependent phenomena.
The researcher considered the CUTGBS to be a ‘figuration’ of a complex co-adaptive and
organic system. This position contrasts with most positions adopted in quantitative research
that consider the phenomenon under study rather to be a substantive and static entity.
I prefer to call the persons who took part in the Validation Study contributors than
participants, mainly because to me their input was more about ‘theory building’ over and
above generating an understanding of events, processes and issues within the CUTGBS. For
this reason the sample for the validation consisted of members from the original sample (CR-
1; CR-2; CR-3; CR-4) a Quality Assurance manager in a Parastatal (CR-5), and a Quality
Assurance Director in a private enterprise (CR-6). These are contributors (participants)
whose comments will be specifically referred to in chapter 6. Some of the contributors had
asked for a debriefing on the results and had provided expansive expositions on the model,
185
university management and comments on the research findings. I dropped copies of my
findings with each contributor with a request that they run over the findings in preparation
for the discussions. See Appendix 14 for the general question. Three focus group sessions
were run in Harare, with CR-I, CR-2, CR-3, at one instance CR-1, CR-4, CR-5, CR-6 at the
other instances and with CR-2 and CR-6. I had further to these focus groups, one-on-one
discussions with some of these same contributors.
Letter from researcher to CUT registrar seeking permission to undertake the research
Letter from registrar to researcher for the director of the CUTGBS authorising the
researcher to undertake research
Consent letter from researcher to participants / respondents.
186
Nearly every research decision and action has ethical connotations and how skilfully
researchers navigate their way through is a function of their experience, reflexivity and
principled standing (Jones, Torres and Arminio, 2006:154). Issues of ethics in research relate
to research design, sampling, researcher-participant relationship, interpretation and handling
of findings, presentation of findings, language and communicative patterns. Throughout the
project the researcher avoided the use of marginalising and devaluing language and always
addressed respondents with dignity, respect and honour. The study has tried to deliver on the
promises communicated implicitly and explicitly in the statement of purpose and to cover as
completely as possible the research questions. The protracted effort to complete this project
is part of an intrinsically felt moral-ethical obligation owed to family, the University of
Pretoria and all, including respondents, who gave their precious time to provide input into
this research.
There was need to display genuine sensitivity to the manner in which participants received
and perceived interview questions and the researcher’s presence in their space. In a
voyeuristic search for good stories and narratives the researcher avoided looking too pressing
on respondents. Participants were allowed to flow, undomesticated in their experiences, their
talk and their reflections. During questioning the researcher avoided looking overly pestering
yet kept looking for ways to get detailed descriptions of narrations. Further to ensuring rigour
in the analytic process, this document provides full evidence of the claims it makes. The data
gathering process proceeded over a year with lots of loops, cycles, and work-backs, all in an
attempt to attain high levels of rigour. The long period facilitated a fuller embedment in the
CUTGBS, and in establishing intimate relations with many who contributed to the research.
Jones, Torres and Arminio (2006) say of the long interaction between researcher-
phenomenon-participants “this is the truth-telling dimension in the ethics of qualitative
research” and they further argue that “the researcher must spend significant time with the
data and the analytic process to get the interpretation of the data close to participants’
meaning”. Long period and researcher-researched relationship both help to co-construct the
truth.
187
Sections of the document that make special reference to the NAMACO, ZimCHE, the CUT
and the CUTBS were proofread by senior officials in each of these institutions. Apart from
contributions from supervisors, many have read and commented on the document. This was
to seek a multiple perspective on almost each key aspect of the study. Jones, Torres and
Arminio (2006:171) say that it is ethical to have several other people read, proof and
comment on one’s research. An ‘inquiry auditor’ went through 50 per cent of randomly
selected transcripts to check on the congruence between what was written against transcripts
generated from the research.
4.10 Conclusion
This chapter discussed the appropriateness of the qualitative research and case study
approach. A triangulation approach was adopted to thicken the descriptions of aspects of the
‘case’. Plugging data from one-on-one interviews, focus groups, document analysis,
observations and literature densified understanding of the ‘case’ and strengthened the value
of the findings in theory building as in shaping policy and practice. A Validation Study was
conducted with the purpose of ensuring that the renditions by participants were adequately
represented and truthfully interpreted and proportionately plugged into literature, theory and
each other as primary data. The Validation Study thus served as an ethical exercise, a means
of improving the generalisation of findings, validating research conclusions and particularly
the potential applicability of some four ‘models’ herein proposed. The next chapter discusses
findings from the study.
188
CHAPTER 5 - FINDINGS AND THEIR ANALYSIS
5.1 Introduction
In this chapter findings gathered through observations, one-on-one interviews, focus groups
and document analyses are presented and analysed. I begin by discussing the transformations
in the higher education landscape. From here we set the basis for evaluating the sense in
trying out a complex model like QFD. In Section 5.3 QFD is discussed from three poles – the
CUTGBS, the academia and the practitioner-consultant points of view. Such a multilateral
analysis should highlight the common, the peculiar and that which was specific of QFD to
the CUTGBS: what was QFD? Why QFD now? How did they do the QFD? The latter two
questions are subjects of analysis in Section 5.4. The best way to approach an assessment of
how a multi-stage model was implemented is to discuss the depth and frequency of use of
each stage of the model – how the model was dismembered and re-membered over time.
After all quality assessment is a step-by-step process and not a broad-spectrum development
(Williams, 2011:2; Veiga, 2011). This implies an in-depth analysis of the profundity and
frequency (intensity) of use of techniques and tools pertaining to that particular stage
(Altbach and Salmi, 2011:138). Such an approach should preserve the unity of the model,
thus affording us a fair assessment of its adoption and diffusion (Kennedy and Fiss, 2009).
A stage-by-stage analysis affords the study an opportunity for differential and poignant
appraisal of what the stage meant to the CUTGBS as well as why it was treated the way it
was. Such specifics would dissolve, vanish and lose their taste and visibility in a blanket and
generalised discussion of the adoption of a model. This is the focus of Section 5.5 which will
form a fairly long section because it covers the various stages of the QFD model. Section 5.6
analyses results on the variable use of the many QFD tools and techniques - a measure of the
amount of permittivity of QFD in the CUTGBS and its adsorption and assimilation into the
quotidian life of the processes in the CUTGBS. In Section 5.7 staff perception of the efficacy
of QFD as a programme quality assurance model is analysed. The chapter closes with what
management has done in response to the application of QFD in the CUTGBS. This is the
domain of Section 5.8. From such results we can infer whether management felt QFD was a
worthwhile assert or investment or the attempt to adopt it was just but another fad or game.
189
5.2 The higher education landscape today: a literature review
Massification, commoditisation, marketisation and increasing diversification of students,
curricular and modes of delivery are the trends in today’s higher education delivery system.
Altbach (2012) refers to the MacDonaldisation and commoditisation of higher education, a
phenomenon in which learning centres are being franchised to host programmes from other
universities. All these phenomena point to an increase in the intensity of competition in the
higher education arena. Technology has also almost completely removed the need to be
physically present in front of a human teacher by digitalising classrooms onto iPhones and
smartphones (Cochrane, 2014). Technology has been exploited to create a new generation of
ubiquitous learners or students. Business schools have also joined the leagues of well-paying
employers, as other universities and business schools have gone on the stock market
(Altbach, 2009). Other business schools are treated as Separate Business Ventures by their
mother universities. Postgraduate students in Zimbabwe pay their full tuition unless they are
sponsored.
Selling education has become acceptable ‘big business’ attracting many players that include
private for-profit, private non-profit, church-owned, government or state higher education
institutions (HEIs) and offshore branches of foreign institutions. Equally so, the number of
garage universities (World Bank Report, 2000:32), pseudo universities (Altbach et al.,
2001:8, 2007, 2009), and demand-absorbing (Lin et al., 2005; Bing, 2009; Wu, 2009), low-
quality HEIs as well as good-quality HEIs has increased. Quality literacy is also growing. A
director with the ZimCHE [DR1] said the regulator was experiencing a surge in the number
of inquiries on the authenticity of higher education institutions and some programmes.
Despite lots of research, the cleft between expectations, standards, and measures between
external quality management (EQM) and internal quality management (IQM), continue
unabated (Taylor et al., 1998; Newton, 2000, 2012; Gallagher, 2010; Ramirez 2013;
Vukasovic, 2014:45). A senior personnel in the Zimbabwe Ministry of Higher Education and
Training explained that the focus of the ministry was to ensure that academic standards were
lifted beyond a threshold and that it will be left to the perceptions of individuals that
university X is better than university Y. Due to the stiff competition, universities were
adopting different models of management, curriculum design and delivery modes with the
190
aim of increasing their competitive standing. The most common issue sought in building a
competitive advantage was the perceived quality of offerings. HEIs are engaging in
advertisement campaigns and image-management activities to increase their visibility to the
market of prospective students, employers and sponsors. Ambush marketing strategies,
market signalling and adoption of other marketing behaviours based on Optimum
Distinctiveness and Strategic Balancing have become part of the competitive game of
institutional and programme posturing.
The CUTGBS, the subject of this study, adopted a QFD-based model in a structured M.Sc.
Programme launched in 2005. A synthesis of CUTGBS staff comments indicates that their
idea of the QFD was to built internal strategic capabilities for defining quality in terms of
market dynamics; develop tools and resources for creating, managing and delivering that
quality, and create a competitive advantage based on organisational learning. The diffusion,
called for or unintended, of typical business-world behaviours and concepts into the higher
education terrain has gone up exponentially in the last few years as explained in Chapter 1. It
behoves and requires on us therefore that when we deal with the migration, by exportation
(push) or importation (pull) of the business jargon or philosophy we examine these most
critically.
In Section 5.3 below I discuss the definitional characteristics of QFD from an academic
literature point of view and compare this with the points of view on the blogs of the
practitioners and with the expressions of CUTGBS staff. Such a comparative analysis is
called for given the record of inconsistencies between scholarship and the practice of quality
(Ramirez, 2013:127). Even within the practice of quality the discrepancy between offered
quality and perceived quality and between perceived quality and expected quality continue to
haunt higher education (Altbach, 2009; Lomas, 2007). Similar concerns of discrepancies
between hypothesised quality and planned quality; between planned quality and offered
quality; between offered quality and expected quality, and between marketed and perceived
quality were echoed among students, academics, ZimCHE staff and other participants.
In summary the higher education landscape has a rough topography with lots of disconnects
191
yet offering new opportunities for scholarship, new modes of delivery, the challenges of
quality and of transformations and real collaborations among all stakeholders. For more
disconnects see Roberts and MacLennan (2006:31) Bolden et al (2010:53), Palmer
(2012:306).
192
In Chapter 2 the use of QFD in education was discussed with all examples provided
indicating overwhelming success. But apparently what each author was referring to as QFD
was but a stage or some techniques that are mainstay to QFD. None of those researches, apart
from claiming examining use of QFD ever used all the stages of the QFD framework or
model. This study focused on the definitional aspect of QFD persuaded by the fact that
definitions matter in that they influence people’s relation with the action that relates to that
definition (Vagneur, 2008).
Secondly there are normally variations between scholarly perception and practice (Ramirez
and Berger, 2014; Kleijnen et al., 2014). In answering the research question ‘what is the
nature of QFD? I had three poles in mind: scholarly perceptions of QFD; consultant-
practitioners’ perceptions and the perceptions held in the CUTGBS. From all three poles I
found some variation in emphasis of aspects of their perceptions of QFD. I got scholarly
perspectives from literature review and for the QFD consultant-practitioners’ perspectives I
visited many QFD consultant-practitioners’ blogs. I synthesised participants’ perceptions of
QFD from their explanations and narrations. It would not have helped other parts of my
research to ask participants for a cut-and-dried definition of QFD. What I wanted from
participants was what QFD was to them and how they were translating it in their day-to-day
work-life. There was no variation in the broader understanding of QFD by scholars,
practitioners and the CUTGBS staff. There was again no variation in why QFD should be
implemented.
Answers from all three poles of my research pointed to the need for alignment, integration,
customer satisfaction and revenue generation. An arresting observation was the way
CUTGBS staff were articulate with market issues, management issues and the QFD lexicon:
they had a strong focus on the pragmatic meaning of QFD and its implications. Sources
however varied in emphasis of what QFD was. The ‘theory’ of the communication matrix
explains variations in emphasis during communication in terms of a trade-off among
purpose, assessment, audience and format. Figure 5.1 attempts to explain this phenomenon.
193
Both scholarly and blog sources defined QFD as:
a philosophy (Mukaddes et al, 2010; Sallis, 2012);
a system (Jnanesh and Hebbar, 2008; Hafeez and Mazouz, 2011:33);
an approach (Akao, 1990; Crow, 2014);
a methodology (Bosch and Enriquez, 2005:30; Ahmed, 2006:193; Najafabadi et al.,
2008:26; Gonzalez et al., 2003, 2008:38; Paryani et al, 2010:1; Sallis, 2012);
a method (Akao, 1994:339; Sahney et al., 2004b; Jnanesh and Hebbar, 2008; Quinn et
al., 2009);
a process (Quinn et al., 2009; Paryani et al, 2010:1);
a tool (Karsak et al., 2002; Jnanesh and Hebbar, 2008; Mukaddes et al, 2010; Hafeez
and Mazouz, 2011:33); and
a technique (Singh et al., 2008:162; Jnanesh and Hebbar, 2008:1; Kay, 2008:501;
Zailani and Sasthriyar, 2011:128) and a means. Figures 5.1 and 5.2 indicate all these
characteristics: variable emphasis and the definitional attributes respectively.
In Chapter 2 the thesis discussed the Theory of Constraint and the 14 Excellence principles
which are basically the way people think organisations create their failures or their successes.
Chapter 3 discussed assumptions around what quality is and quality assurance work can be
194
run. In chapter 6 a point is made that to improve quality performance at either the global or
the narrower levels of a course or programme it is important to examine and change the
epistemology, the mental models and philosophies that created the strategies, methodologies
and models that are the hard-to-see causes of the underperformance. In viewing QFD as a
methodology the implication is that QFD should be looked at from differing scopes of
complexity. At a more sophisticated level QFD is a system of interweaving methods maybe
operating with different stages of a model or strategy.
Each method will be enacted through a system of processes, procedures, tools and
techniques. This view is evinced in Chapter 3 when the study discussed the different
techniques and tools used at each of the stages of the QFD model. My research participants
used the terms in Figure 5.2 in their reference to QFD and what was happening in the
195
CUTGBS. SM1 once referred to QFD as an approach then as a philosophy. AC3 and SM5
both talked of QFD as a method and as a system. This double ‘meaning’ is frequent in
literature. Jnanesh and Hebbar (2008) in the same document refer to QFD as a system, a
method, a technique and a tool. Mukaddes et al, (2010) in the same document also refer to
QFD as both a philosophy and a tool. Sallis (2012) refer to QFD as both a philosophy and a
methodology in the same document. Akao (1994) similarly refer to QFD as an approach and
later in the same document as a method. Hafeez and Mazouz (2011) writing in the same
document refer to QFD as a system and again as a tool. Even where no explicit use of the
term ‘‘QFD’’ was made, participants were using the terms ‘process’, ‘method’, ‘system’,
‘philosophy’ and means quite frequently and effortlessly. Participants also used the term
‘total quality management’ in reference to the strategy of ensuring production and delivery of
competitive products and services in the CUTGBS.
A more profound analysis of the definitions suggested that the two key words are philosophy
and methodology and the rest were perceived as aspects of these omnibus-like terms. QFD as
a philosophy apparently subsumed the concepts of ‘approach’ and ‘system’. QFD as a
methodology apparently embraced aspects such as ‘method’, ‘process’, ‘tool’, ‘technique’
and means. Whatever definition a participant subscribed to, two aspects remained
conspicuously clear in the narratives: that QFD was about a way of thinking, a mental model
and that QFD was also about a practical path of doing or achieving something through some
pragmatic action. I call this phenomenon the dual nature (duality) of QFD. Retroflection to
Chapters 2 and 3 show how a methodological (technical) approach to quality and quality
assurance is impotent. The inverse is equally true.
A purely philosophical (rational) approach to quality and quality assurance is barren. The
base-line observation of the study is that philosophy and methodology are mutually inclusive
and where they are managed creatively they have the potentiality of an ‘enhancement’ or the
1+2 = 5 effect. This phenomenon is not uncommon in organisations and it is called the
‘synergistic effect’. In plant photosynthesis it is referred to as the ‘Emmerson Enhancement
Effect’. The implication of the duality approach is that a quality approach that synergises the
philosophical and methodological perspective of QFD will have an aggregate rendition
196
bigger than would be gotten from the two operated separately. Indeed to beat the challenges
for quality and quality assurance in higher education it is necessary to harmonise HEIs’
formal design infrastructures with their emergent designs (the way that people naturally
redesign HEIs as part of their actor-hood and evolution). Monolithic models cannot run the
quality show solo. Figure 5.3 below indicates the philosophy-methodology duality of QFD.
Figure 5.3: The dual nature of QFD and its relationship to market orientation and the
organisation’s competitive advantage as synthesised from literature and narratives
So whether QFD was referred to as one or the other, it was about emphasis and the inherent
trade-off among the elements of communication (purpose, format, assessment and audience)
that narrators make during discourse. The duality, inherent in QFD, creates its unparalleled
efficacy as a model or framework for improving the competitive performance of an
organisation. Its philosophical nature informs the management strategic orientation while its
methodological nature informs the products and services strategies. I reiterate this point and
exemplify it in Figure 5.3 above. By interweaving a management strategy and a products and
197
services strategy those using QFD should be able to align management to customer needs and
wants. A QFD approach helps the adopting organisation to take cognisance of the often
taken-for-granted cultural, policy, practice and contextual assumptions in ways that other
quality models don’t. Research has found that these nuances define the success or fall of a
model (Blackmur, 2007; Law, 2010; Ramirez, 2013). Therefore to be fair to those whose data
was used in this thesis, I took the broader understanding of QFD as both a philosophy and a
methodology. Literature and the voice of my participants portrayed QFD as a vehicle for
improvement and innovation. I refer to this as the quality improvement-innovation helix.
Like the DNA in living organisms this helix denotes the genome of never-ending quality
improvement which the Japanese fondly call kaizen.
In summary, QFD is what become when a philosophy and a methodology mutually embed in
each other to offer their tools and techniques an enhanced synergistic effect that allows the
resultant QFD to deliver to the satisfaction and delight of the customer. It is like a jet that can
take any orientation without dismembering. The pragmatic value of QFD is that HEIs can use
the QFD model to pursue their own ‘kind of’ quality as demanded by their own ‘kind of’
customers.
A personnel staff who attended the Retreat said that one of the important achievements of the
sessions was that people came back with new fears, new ways of looking at their
Departments and new dispositions to share ideas and generate resources. These pressures, felt
at the university level were key strategy shapers at the level of the CUTGBS and other
Faculties. A top management level participant said that the reason they had gone for the
199
Retreat was to seek integration and to sharing ideas on how the university could stand and
operate as a strong and coherent organism [SM6]. Apparently some sectors had gone to the
Retreat with held positions to defend. A top management official indicated that they had a
preconceived and shared posturing from which they would influence the amplitude and
direction of discourse at the Retreat. I find this a reasonable strategic behaviour. Whilst
management takes a detached position in corporate discourses of this nature, it is important
however to hold the handle and keep pulse with proceedings. Clashes are not altogether
malice in conditions of deliberative democracy (Gutmann and Thompson, 2004). Sevier
(2003), writing on strategy development in universities, described it more as a path to pain
than to plenty (p. 18). This was also the first massive and most inclusive Strategy Planning
session the Chinhoyi University of Technology had ever had.
The other strategic issues that emerged at the Conference Centre included the need to
generate funding; producing five-year and ten-year projections of programmes and faculties;
designing market-oriented curricular; having strategic plans on quality improvement at
department and course levels; and the need for staff to self-develop and seek tenure. One of
the key challenges was that the main source of funding for the departments was fees and the
bigger the enrolment the more income the Department would have. However, increases in
enrolments were becoming a strain on resources and there was a feeling that this was pulling
the quality and reputation of the institution down.
In a nutshell, QFD was not chosen but became the emergent epitome of a response to the
many strategic issues that arose in a seriatim of self and context analyses. From the above
description of the interplay between external and internal forces for a new model for doing
business I turn to a discussion of the pull and push factors for adoption of QFD in the
CUTGBS.
Figure 5.4: Organisational challenges faced by the CUTGBS as synthesised from data
(S)he also said that understanding the context and aligning the programme to the broader
trends and the particularities of the university context was a strong strategic aspiration.
(S)he explained that finding a competitive niche for the M.Sc. Programme and the CUTGBS
were critical pursuits. The factors that were continuously discussed in conversations on
201
strategy choice were about legitimacy, at least in terms of meeting the ‘national skills and
knowledge priorities’; profitability; marketing and responses to the international trends in the
provision of higher education. From the conversations the thread of thinking that was most
salient was that if the war on quality is won, then the battle for legitimacy will be simpler to
fight. One phrase that was most repeated was ‘profit or perish’, suggesting that everything
that would guarantee a good stream of enrolments would matter a lot to the CUTGBS. One
academic stated it bluntly that their lives and careers were tied to the profitability of the
CUTGBS. The phrases ‘profit or perish’ and ‘if you can’t make it, close shop’ were often
repeated by participants. This could be an indication of the ruthlessness of the master’s
degree market.
The closure of departments and courses is not unknown in the history of higher education.
Alvesson and Skoldberg (2009:288) once observed that we can garner rich and interesting
insight into the values people carry in their doing of something and commitment to
something by how they ‘‘show their understanding of the driving forces and social
conventions as regards modes of expression and legitimate motives’’. A senior manager at
the Chinhoyi University of Technology said there was no pressuring of Departments to make
particular choices of models or strategies [SM3] and the choices of what strategies to adopt
was quite free and open [SM5].
The question of why and how organisations adopt new strategies and models has been
addressed by Abrahamson (1991), Davis and Greve (1997), Palmer, Jennings and Zhou
(1993) and Jones (2014). Kennedy and Fiss (2009:897) say that adoptions may take place
along two vectors. One of such vectors is the rational actor model that is based on the logic
of economic gains. The second is the social legitimacy model which hinges on the logic of
legitimacy. The economic focus was well ventilated in the conversations with the assertion
that the QFD approach was helping in cutting costs and in improving relations with
stakeholders and enhancing market-orientation of the institution and its products and
services. Indeed one of the CUTGBS objectives was to increase the 2012 market share by 20
%. The CUTGBS sought legitimisation by heavily pursuing Voice of the Customer and
linking with outside stakeholders and powerful constituents [AC1]. These tactics for seeking
202
legitimacy are cited by Houston and Paewai (2013:265). The logic is that once there is wide
support then everything will fall in place and with the rational model the assumption is that
once the organisation is efficient and economically strong everything will fall in place.
Actually the CUTGBS sought more than legitimisation and economic gains. The CUTGBS
sought to brand itself with a uniquely high level of optimal distinctiveness on the market. The
following five statements help to show the penchant desire to pursue a strategy or model that
focus on quality and stakeholder satisfaction and also on image, reputation and impression
management:
In QFD contexts the gemba is the actual place where the customer uses the product.
However the Strategy Balancing aspect of this mode of Voice of the Customer in terms of
optimal distinctiveness pursuits and products and services improvements cannot be separated
one from the other. In the perspective of Strategy Balancing Theory the market of employers
and students was showing some sensitivity to the institutions’ resources including strategies
by which they were known to deal with many of their management and products and services
issues. See Goleman and Senge (2014) as they reiterate Deephouse (1999), Brickson (2005)
203
and Glynn (2008) in very profound perspectives on how people and institutions create
identities of self with relation to their inner and outer worlds. Profound changes require lots
of systems thinking; networks that enhance the compound value of external competences;
triple-focus; deep learning and an infrastructure for overcoming our own learning disabilities;
top management’s earnest desire to change the ‘we support’ mentality and bluff to deep
commitment in activities initiated at lower altitudes of the organisation.
In a summative form, because QFD arose from a datum of concerns as the most appropriate
bundle of responses to CUTGBS’s strategic concerns it was charged with the expectations of
meeting the desire for image management and of creating the much needed critical success
factors in terms of economic, financial, management, competitive and quality performativity.
5.5 Response of staff to QFD and the institutionalisation of the QFD model
5.5.1 Introduction
This section answers the research question: How did staff respond to the QFD model and its
institutionalisation in the CUTGBS? Therefore the purpose of this section is to describe
findings relating to how staff behaved with respect to requirements of each of the stages of
QFD as narrated in Section 3.17 of Chapter 3 and depicted in Figure 5.5 below. The
description will cover how staff in the CUTGBS responded to QFD by assessing how they
planned and organised the execution of each of the QFD’s milestone stages. An adopter of a
model may respond to the model by selective implementation or execution of its stages, by
superficial execution of every of its stages, or by first a blanket or aggregate and superficial
adoption followed by an incremental and in-depth institutionalisation of the whole model.
Another form of response to a model is to hybridise it with one or two other models with
which the implementer is already familiar and in which he has built up sizeable trust.
Outright rejection and jettisoning of a model soon after its implementation can be classified
as a response to the model.
To fairly assess the depth of response to QFD one should analyse what staff did, and how
they did it and how their way differed or followed that stipulated in the model. The CUTGBS
adopted all stages of the QFD but was putting more emphasis in some stages of QFD than
204
others. The emphasis also varied in terms of the intensity of use of the QFD tools. The
presence of teamwork was evident and a process of continuously obtaining Voice of the
Customer was equally present. There was evidence of a strong effort to shift the locus of
quality assurance from the external quality management infrastructure into the realms of the
internal systems of the CUTGBS. Thus QFD was becoming an important and effective tool
for strategic quality planning purposes. Figure 5.5 below should help in understanding the
model of QFD the CUTGBS staff worked through.
205
Figure 5.5: Stages of the QFD adopted in the study (Franceschini, 2002; Ficalora and
Cohen, 2009)
206
In the preceding two sections I discussed the first two aspects of the QFD model. Participants
clearly articulated the contextual influence of the higher education landscape and that of the
Chinhoyi University of Technology on a deep-going self-assessment on the part of the
CUTGBS. They further indicated that QFD emerged as the model of best-fit and that
preparations for its adoption and diffusion were an intricate part of the organisation’s life.
Indeed a strategic approach to the adoption of QFD was followed. The QFD model or what
may be understood as a strategic framework arose out of an on-going analysis of both the
internal and the external environments of the CUTGBS.
Strategic frameworks that are imposed or chosen by a vote are less likely to generate
sufficient support and commitment from either the technical or ground level or from the top-
level management. They may also soon lose favour with the organisation as they often show
incongruence with many aspects of the context of the organisation as a result of people
withholding their support. People commit more and give value to what they have chosen
than to those things imposed on them.
The core QFD team in the CUTGBS was running some strategic, planning, organising,
staffing, monitoring and other work responsibilities. However the team did not have
complete locus on these leadership-management facets. The CUTGBS team was apparently
pushing and pulling for greater autonomy from the mother university. The shifting of more
power to the CUTGBS and the increasing prominence of local leadership at the level of the
core team was having far-reaching implications. The professoriate were taking more charge
on leading and planning particularly with regard strategy formulation. The second was
leading and controlling as a means of legitimising the M.Sc. Programme and the CUTGBS as
an institution. The third was on the connection between leading and motivating as a means of
accomplishing staff stability, buy-ins and retention. AC2 explained that the CUTGBS staff
was actually setting an example, that they were determining their frame of work and
conditions of service.
In Section 5.1 I discussed the dual nature of QFD. It was the responsibility of the QFD team
to ensure that the products and services strategies tied in seamlessly with the organisation’s
management strategy. What was generally referred to as the CUTGBS ‘core team’ was a
cohort of four senior lecturers and the Director of the CUTGBS, two secretaries and the
CUTGBS administrator. Each of the seven would so naturally refer to themselves as a team
using the term ‘team’ quite naturally in the interview sessions. Part-time lecturers were being
called in to ‘every major event’, ‘particularly meetings at a CUTGBS cost-to-company basis’
[SM1]. However, some of the part-time academics said they were not able to attend all
CUTGBS meetings and sessions as they would normally have engagements at their places of
employment. There was no other person from outside of the CUTGBS that was part of the
208
core or steering team (2013, March). The cross-functionality nature of the team was basically
their different areas of academic specialisation. The team took the local leadership role,
acting as the key strategy formulation unity in the CUTGBS. Each of the team members was
elaborate on what the CUTGBS planned with regard to management issues and the design
and delivery of products and services. The core team consisted of persons with diverse
backgrounds ranging from industry, educational management, lecturing, and some were
engaged in PhD studies. Despite members expressing a felt sense of benefiting from each
other’s experiences I do not think experiences should be taken in lieu of the actual
membership of persons in the cross-functional team.
CUTGBS staff varied age-wise with the youngest around 32 in 2013 and others in their 50s.
Their exposure to research varied, with some having published more than three articles and
others not. There were promises that each member of the core-team would continue as
chairperson-in-waiting for the future programmes that were on the verge of being launched.
The advantage of such an arrangement is that the chairs-in-waiting would have long
preparation periods. The disadvantage is that the decision forfeits the opportunity of
appointing to chairpersonship people of superior practical experience and abilities should one
appear in future staff compliments.
Literature on teamwork and its dynamics alludes that teams evolve over time, passing
through a number of stages. The orientation in the development is basically from political,
socio-oriented concerns toward task-based concerns. In the research it appeared that the team
had passed the stages of ‘who is who here’ and sounded like they enjoyed cooperation,
communication and collaboration among them. One thing that would come across quickly
and quite strongly was the task-focusing of the staff and the trust in achieving what they
wanted.
To establish all facets of teamwork culture I believe one would need to do more than just ask
people but embed in with the members. The research however limited itself to inferences
from which I can say participants felt there was good link and coordination among CUTGBS
209
staff and between the CUTGBS and its key stakeholders, including committees at the mother
university.
‘‘We work together, how could we not, when we belong in the same university? But it
doesn’t remove that we differ in the way we would like some things done.’’ [AC2]
This was corroborated by two other academics, [AC3] and [AC1], who felt that much could
be done to achieve coordination on things that matter the most, particularly when it came to
working with teams outside of the CUTGBS. The participants did not however feel that the
current level of coordination was detrimental to anything.
Balance of member contribution. Within the CUTGBS core team staff felt there was respect
among members and that ideas could be brought forward and attended to by the team. There
was a feeling that the CUTGBS needed to move more toward totally inclusive behaviours.
The balance of member contribution is about allowing each other space to talk and to display
expertise. AC3 referred to some tendency to block-out on others and not show total inclusion
behaviours. He worried that there were apparent tendency toward groupthink and that
sometimes ‘‘undiscussible’’ issues were quickly glossed over.
Mutual support, Team effort and Cohesion among team members were felt to be present
within the team. Structurally, most participants felt that a greater level of autonomy would
work best for the CUTGBS and for its relations with the mother university. The situation by
2013 was that most decisions would be tabled at forums and committees that had
membership outside of the CUTGBS. CUTGBS staff felt that this modus operandi made the
institution more sluggish and was not in line with the CUTGBS’s principle of ‘being time
based’.
“... we try to be a time-based Business School, and we take time, every second as a
critical resource. But we find it (the time-based principle) not working most of the time
because our resources are controlled elsewhere where time is not important, we see things
differently.” [AC3]
The reason for the much-felt friction was the lack of understanding of how the CUTGBS was
setting standards and modes of doing things. Top university management was felt to be rather
slow and ‘traditional’ for the current dynamics and competition’ faced by business schools
210
today. Whilst there appeared to be much stronger strategic cohesion within the CUTGBS,
strategic bundling across the university was not felt to be visible enough. Staff in the
CUTGBS said being beholden to the mother university was not perfectly favourable to their
hopes of adopting a QFD-based mode of operation.
In a nutshell the core team of academics in the CUTGBS was the structural-functional
epitome of a QFD team. Thus it embodied boundary-spanning roles, management
responsibilities and the responsibilities that would normally fall within the charges of a
Quality Cycle or a QFD team. This duality creates huge opportunities and yet challenges as
well. Both should be attractive to new study of QFD adoptions. The CUTGBS QFD team
coordinated with other external institutions.
However, a staff member says that the main problem in the CUTGBS-student interface was
that the M.Sc. Programme was ‘microwaving’ the student. As a result, students were ‘late’
with everything ‘like fees are outstanding’, ‘there is no communication with the supervisor’...
‘assignment issues’ and ‘a host of these subtleties’. There was concurrence from academics,
the non-academic staff and some students that the structure of the M.Sc. Programme exerted
‘unbearable pressure’ on the student.
“...things have changed greatly over these years, we now have meetings with students at
the beginning of the semester, we run course outlines and they feedback ... we again
conference with them at the end of the semester and we discuss the throughput, the
processes and arising issues. We get their feedback per course and per lecturer, and these
are deposited in the database and we act on that data.” [AC4]
This quotation gives evidence as to how deeply and sophisticated the mentality of continuous
improvement had sunk in with the staff in the CUTGBS. A perception that even the
apparently small issues and changes are of value in enhancing Customer Satisfaction
Performance is evinced here.
212
Another academic had this to say:
“We take the students and we say these are the options on board, they discuss these and
we hear and note their comments .... we get some options that don’t get takers because
they are felt to be not so useful, we take it that it is their choice.’’[AC6]
This quotation exemplifies the responsive nature of the CUTGBS and that they perceived
Quality as fitness for a purpose defined by the student herself as their primary customer. A
Student Evaluation of Teaching Effectiveness (SETE) instrument was used. Student
comments were said to be more on what they liked and disliked about the teaching and not
about what they actually wanted taught. Some students had this to say:
“... yes we may know what we want taught but who do we tell, when and do we wanna
get that in the end? We are here for a very short time for each course (subject) and we are
more worried about getting it through and we move on to the next thing.” [FG1-ST3]
Another lecturer said they get comments from students in an opportunistic way.
“We try to be very open with them. It helps in making them feel free to talk to us. These
are adults and this has two advantages: they don’t just come with everything like high
school kids would, yet eeeh they also have the temerity to approach you and talk to you
like an adult. Most of the issues that the students bring relate to studies per se, things like
being unable to attend one or so lectures and concerns about the research component of
the M.Sc. Programme.” [AC3]
“... it’s quite difficult, here are two parties wanting to do or are actually doing a
transgression. Both are taking care not to be discovered. It’s difficult actually to discover
them. We can’t say some transgressions are not happening, we can only say we haven’t
discovered this or that. Maybe that we don’t have any may mean that none of these
transgressions is happening. We have a very sophisticated system of sniffing, doing
surveillance and catching on anything once we get a hint.”
213
The tool used for Student Evaluation of Teaching Effectiveness consisted of various criteria
and items. Students were using the Student Evaluation of Teaching Effectiveness instrument
to express their sentiments and to appraise the lecturer. However literature says that further to
these two functions, students can also use Student Evaluation of Teaching Effectiveness
instruments to punish lecturers. None of the three focus groups admitted to having used the
Student Evaluation of Teaching Effectiveness instrument in punishing the lecturers, but they
thought it would happen. I inquired on what type of punishment and how the students
thought the lecturer would actually get punished?
“... you see if one does not pitch up for his slots then suddenly brings in a difficult
test.” [FG1-ST4]
And
“... one is so niggard with the marks.” [FG2-ST2]
And
“Yaa we have these rude professors with lose stuff you can’t tell where you are
from and where to with the course then suddenly there is a difficult paper with
stuff that never was in the lectures.” [FG4-ST1]
I find two issues in these comments. One, that the students demand value for their money,
and that they really show an interest in receiving good quality instruction. The second is their
potential to (ab)use the Student Evaluation of Teaching Effectiveness as a stick to ramrod
those who can stand in their way of progression, either for good reasons or for bad ones. I
researched on issues of mark allocation and found that markers can be consistent, too lenient
and actually niggard. Aware of these ‘mishaps’, institutions allow for internal, cluster or
external moderation of a sample of script papers. The fact that students can settle felt scores
with lecturers is not uncommon in literature. Anderson (2006) said student surveys could be
missing their purpose unless the student appreciates how they could benefit from these
surveys. It is important that management must factor this undesirable potentiality. One way is
to design survey instruments that focus on issues that have value to the teaching-learning
interactions and to consider timing of the Student Evaluation of Teaching Effectiveness.
Other students indicated that they tried to be fair in their evaluations. Asked what ‘fair’ looks
like, the impression was that lecturers who are ‘good’ needed to be rewarded with good
214
ratings. One student [FG1-ST4] expressed ‘goodness’ as meaning being mature and
understanding the complexity of being a student and a worker and a parent and maybe a
patient of some illness. Lecturers’ expectations were that students would rate them fairly, but
they did not repudiate the possibility that students could settle felt scores through the Student
Evaluation of Teaching Effectiveness process. Lecturers felt that student comments mattered
in that they were input for their own self-assessment and as part of Voice of the Customer.
Management was using the Student Evaluation of Teaching Effectiveness as input in subject
allocation decisions which mattered a lot to the lecturers because it was a source of income to
them [AC6]. Both part-timers and full-time academics are paid an hourly rate over and above
their salaries for teaching in the M.Sc. Programme.
In a conversation with a lecturer in the comparative case there were similar feelings about the
potential abuse of Student Evaluation of Teaching Effectiveness by students. He explained
that it would be easier for lecturers of some courses to show lenience than lecturers of
courses like Statistics and Financial Accounting. Interestingly, CUTGBS students did not
link high ratings with rigour and strictness on the part of the lecturer. However, the Director
of the CUTGBS was quick to say that he expected that students would answer Student
Evaluation of Teaching Effectiveness questions accurately and without fear or favour. He
admitted that the CUTGBS took the Student Evaluation of Teaching Effectiveness ‘quite
seriously’ and that they did them because they believed in their value as part of their Voice of
the Student.
In Chapter 3, the Voice of Customer was discussed and the point made was that Student
Evaluation of Teaching Effectiveness are only valuable to the extent that they are
appropriately designed, that they get useful data about the student and that the data can be
processed and factored into institutional strategies and decisions. Anderson (2006:166) says
of Student Evaluation of Teaching Effectiveness that they are a very powerful tool, but
unfortunately in the hands of those with marginal understanding of how they should work.
I observed from the conversations that both parties can manipulate Student Evaluation of
Teaching Effectiveness instruments. The degree to which academics are likely to manipulate
215
Student Evaluation of Teaching Effectiveness is influenced by the clout of the completed
Student Evaluation of Teaching Effectiveness to influence distribution of office (promotion
or tenure decisions) and social benefits like recognition and standing among colleagues or the
number of assigned courses. The latter becomes of interest, particularly in instances where
lecturers are paid over and above their salaries for teaching what they are employed to teach
and the allocation of subjects is based on previous Student Evaluation of Teaching
Effectiveness ratings.
Voices of Business
Voice of Business in the context of QFD may refer to the business concerns of the
organisation. For self-supporting, for-profit institutions or programmes it is important to
consider the institution’s desires to create a profit. However, Voice of Business may also
refer to the voices in education of the various experts and connoisseurs whose inputs can
improve (or destroy) the quality of education delivery. Voice of Business was gathered
mainly from external examiners who were hired each semester and the many individuals who
participated in the various review teams that did audit and assessment work in the CUTGBS
and the mother university. However, there were no opportunities for the external examiner to
directly discuss with the concerned subject lecturer either before or after the moderation
exercise. The dangers with distant indirect interaction are that the lecturer may not always
feel that the comments of the external examiner were based on the peculiarities of the
institution or were tolerant of the academic’s or student’s viewpoint. This is not uncommon
with subjects such as Corporate Governance. At master’s level lecturer practical experience
and his professorial orientation tint their approach to a subject, its didactics and the
instructional-pedagogical values accompanying the teaching-learning interaction.
Most lecturers were active in researching, particularly on content relating to their subjects.
This helped them keep abreast with developments in the higher education landscape. It was
widely assumed that knowledge gained from professional activities like seminars, workshops
‘and all that’ would be ‘ploughed in to fecund the curriculum and the instructional processes’
[SM1]. There was encouragement to focus research on consultancy and on strategies of
developing institutional performance. How much incentive would shift lecturers from
216
researching for their publication records to researching for the benefit of practical knowledge
needs to be explored. There were various agencies exogenous to the CUTGBS whose voices
were being considered in the CUTGBS, for instance, the ZimCHE, the Ministry of Higher n
and Tertiary Education, and the CUT Directorate of Quality Assurance.
Voices of Employees
Voices of employees were gathered in meetings and on performance appraisals. Some staff
felt that there wasn’t a fair space and favourable ambience for them to put forward their
requirements. There was a preference for a shorthand system that is quick to respond and to
understand better their concerns. A respondent explained that the CUTGBS was part of the
university and bound by wider university policy regime, and many of employees’ complaints
or requirements should be channelled through the wider university system. As a growing
institution, the CUTGBS ought to be at pulse with the sentiments off all its workforce. This
would need a robust Research and Development or survey system to constantly gather such
sentiments. However an academics felt that top management was paying attention to their
presence, to their ideas and aspirations about the CUTGBS. [AC1; AC2; AC3;AC4; AC5]
Voices of Market
Every academic staff member in the CUTGBS engaged in collecting ‘Voice of Market’. The
Director of the CUTGBS took it upon himself to go into the industry and get first-hand
information. Senior management personnel [SM1] explained that the main targets in such
exercises were the alumni because:
“... they are the ones who have gone through this course and are best placed to say this I
got it but it does not work, this and that were time wasters, you see, then you can say how
do we improve”? [SM1]
The Chinhoyi University of Technology had the office of the ‘Directorate of Quality
Assurance’, which oversaw quality issues across the whole university. In 2013, the
Directorate was ‘‘... still busy putting up systems in place.’’ [SM2]. The ZimCHE had legal
and professional channels of communicating with the university on issues relating to its
work. The Chinhoyi University of Technology was one of the first universities to establish
217
the office of the Directorate of Quality Assurance and by March 2013 had the highest
compliment of PhD holders among their teaching staff. Despite all the efforts to link and
gather data from various stakeholders, there were some professional institutions that could
help in a number of issues. For instance, the National Manpower Council (NAMACO) has a
wealth of experience in curriculum design and research and working with them should
benefit the CUTGBS. Four of the participants felt NAMACO’s activities were more suiting
to diploma-level studies. A Talent manager who had experience with NAMACO and
university lectureship felt that NAMACO had a huge contribution to degree level education
insisting in that the practical focus of the diploma level education needed to continue through
higher education. (S)he lamented that “... we can boast and say we are a very educated
country, but what does a degree serve when you cannot think contextually, create, and
innovate’’. In the following section the value of listening to both downstream and upstream
audience is spelt out. But there was a concern raised by a student relating to the Voice of the
Customer. (S)he suggested that lecturers teach what they want and are comfortable with so
suggesting anything to them would not make them change their course outline [FG1-ST2 and
FG2-ST3].
Summatively, the student and Industry apparently contested for the key customer position
with other voices being collected from academics and other sources of competences. Voice
of the Customer was being collected mainly by surveys, interviews and focus groups for
purposes of Product Planning Matrix, Customer Satisfaction Performance, and Competitive
218
Satisfaction Performance. More software and technology could be deployed in the collection
as in the processing of the Voice of the Customer. The translation of Voice of the Customer
into management and products and services strategies was evident.
The observation made by QFD gurus like Franceschini (2002) and Ficalora and Cohen
(2009), among others, is that the Product Planning Matrix becomes the repository for
important quantitative data about each customer need. The M.Sc. Programme was based on a
felt need for skilled senior managers [SM-1] and its curriculum was designed and delivered
to meet the senior managers’ needs for examplar leadership performance. The CUTGBS staff
claimed that their curriculum had been validated by Industry and by the student through
Student Evaluation of Teaching Effectiveness over the years [AC-1; AC-3]. However some
felt there was need to re-orient the focus of some concepts and in some instances
contextualise some of the theories [AC-2] because a good number of the students were
indigenous business owners not actually formal employees [AC-1; AC-3]. In this
perspective, the study investigated the intensity with which the curriculum documents, course
outline and assessment records recognised this emergent need from a ‘new’ type of student
wanting to be an employer and services creator.
Institutionalisation of QFD by aligning Voice of the Customer with the M.Sc. programme
and curriculum
The design of the curriculum was basically in the hands of the CUTGBS as a university
department. The proposed curriculum would be escalated to the School Academic Board,
219
which would appraise and then escalate the prototype curriculum to the Senate for appraisal.
The study did not establish that there could have been direct external involvements in the
design of the M.Sc. programme curriculum. In the quixotic sense, contributors in curriculum
development would be the National Manpower Council, the ZimCHE, members from
professional associations, accrediting boards and Instructional Technology development
specialists. It should make good sense to involve academics who lecture at undergraduate and
diploma levels. It could not be established that these had actively and directly participated in
the drawing up of initial curricular or in the validation of the prototype or actual curriculum.
Curriculum mapping should improve both vertical and horizontal integration of content and
perspectives. Well implemented QFD tools could help ensuring that the M.Sc. Programme is
built on the foundations of prior studies and that it enjoyed sufficient internal horizontal
integration. Some students testified that some level of ‘redundancy of content’ [IST2; IST3;
IST4] across some M.Sc. Programme subjects could be felt and too vertical redundancy
whereby undergraduate content was ‘unnecessarily’ consuming instructional time [FG2-
STD3].
Temtime and Mmereki (2010) observed this phenomenon in their study of graduate business
education in Southern Africa. An academic admitted that some duplication can be possible,
particularly with subjects like Strategic Human Resources, Labour Relations and Strategic
Management (AC6). An academic felt that such redundancy was useful in helping the
students realise the inter-linkages among subjects and disciplines. This was consistent with
the findings of Biglan (1973) and Royal, Gilliland and Dent (2014). SM1 and SM5 said that
the CUTGBS ran comparative analysis of course outlines. While content redundancy for
purposes of scaffolding is a desirable deliverable of curriculum design, undesirable
redundancy that wastes student time should be interpreted otherwise, particularly in time-
based organisations. I feel that where content is revisted, it should be within the perspective
of current level thus showing how it is evaluated or applied or analysed at a higher level
using new perspectives rather than sheer repetition.
While the philosophy for curriculum design was perverse the methodology was apparently
not well appreciated. An academic expressed that curriculum planning was one area the
220
higher education sector needed to give lots of attention, yet he lamented that there was very
little interest in learning curriculum designing among university academics [AC5].
Course outlines were designed by each lecturer for his subject, presented to a board at the
CUTGBS and discussed at the School Board level. Some four lecturers said they had adopted
course outlines they had found in operation and that they had, however, included some
modifications. Development of course outlines was somewhat a ‘lone ranger heroic’
endeavour (Senge et al., 2012) which some lecturers defended on the basis that the approach
allowed them to piece together different perspectives and benchmark without ‘distraction’
from fellow lecturers who don’t have the subject expertise. Salmi (2011a:6) says that it is
unrealistic to think that ‘reproducing’ the ‘organic, academic models’ of top-class business
schools ‘is possible’. He goes further to say:
“...it is impractical to envision shopping around and bringing curricular fragments from a
variety of top-notch institutions across different countries and cultures and guessing
everything could easily gel together and fall in place to create an authentic learning and
research culture in the new university.”
221
In the designing of course outlines, most academics were worried about how other lecturers
and students would react to the course outlines and whether there would be sufficient time to
go through the course outlines. The concerns were about depth, scope and length in the light
of customer expectations. Other academics pointed to the heterogeneity of the classes as both
a huge resource and as posing difficulty in piecing up a course outline that would satisfy
every student [AC5]. Some lecturers portrayed a somewhat different picture saying that they
were using Voice of the Customer database to align course outlines with students’
expectations.
The heterogeneity of the class means students brought in a cauldron of expected learning
outcomes and diverse quality expectations. Lecturers were finding it difficult to construct a
one-size-fits-all course outline and instructional experiences [AC6]. This brought in a
discrepancy between the offered quality and the expected quality for some students. Some
students did express some concern about what they had expected and what was actually
being gained. With the use of technology, which allows for the exchange of documents
concurrently, working on them using different coloured ink and merging documents from
different sources, it should be possible for a huge team to work on a curriculum or course
outline from long distances, as virtual teams.
Google+, Twitter, Posterous and Drive are optimised communication channels that can
facilitate this process (Cochrane, 2014:66). The two Six Sigma roadmaps of Six Sigma
Process Design, and Technology for Six Sigma were not being exploited fully within the
CUTGBS. These two roadmaps are discussed in chapter 3, which discusses in depth the
methodological aspects of QFD. One way of covering up for losses due to a broad curriculum
is to allow for individualised out-of-class lecturer-student instruction or mentorship.
However, with large classes this type of interaction is constrained by the ensuing large
demand on lecturer time and by the difficulty of accommodating all the students at any one
time most of the times. These deficiencies highlight the need for new approaches to technical
and pedagogical support through a robust TFSS (Technology For Six Sigma) approach and
the need for sustained interaction and provision of support for communities of practice
through SSPD (Six Sigma Process Design).
222
From the presentations from academics and students there was incremental inclusion of
students’ voices and expectations in the offered curriculum, even in those courses or subjects
the current lecturers had taken over from previous lecturers. New lecturers were making
efforts to accommodate comments and concerns from students as well as those from industry
and own lecturers’ research efforts. An appreciation of the value of Voice of the Customer
and its inclusion in products and services was institutionalising. However, the process
seemed to be taking rather long and was apparently ad hoc in some cases.
Lecturers were encouraged to develop modules for their respective subjects. The Directorate
for Quality Assurance’s website had a guideline on the production of modules and course
outlines. The Academy of Teaching was equally supporting this effort by providing training.
Some subjects already had some kind of module and some lecturers were busy writing
modules for their subjects. However, some felt that they had too large a load to have any time
for writing modules. Two of the lecturers, [AC1] and [AC2], suggested that the University
should grant leave to those engaged in writing and support the process financially. The
current scenario was that lecturers would write a module in their spare time and then be paid
an amount of US$500 (July 2013).
223
In summative form, the Product Planning Matrix (PPM) approach was deployed in the
development of the M.Sc. Programme, course outlines, and other teaching-learning materials.
There was room for exploitation of charts, tables, and other more robust software and
matrices in order to improve the WHAT-HOW alignments and both vertical and horizontal
curriculum alignments.
The theoretical perspective underpinning this study is the Theory of Constraint. Each stage of
the QFD methodology or one or so of their tools can become a major constraint or just a
significant one, which would cause the quality of products and services to suffer. This fact
has been considered with huge concern and throughout chapters 2 and 3. The study
investigated what students rated as the key satisfiers in the M.Sc. programme. The nature of
the M.Sc. programme in terms of the subjects making up the curriculum had high rating.
However, students did not seem to like the duration of the M.Sc. programme. Apparently
many preferred a longer duration with more in-depth learning. One of the CUTGBS staff
described the M.Sc. programme as ‘microwaving’ students. Apparently the duration of the
M.Sc. programme was just the same as similar master’s degrees in other universities abroad
and in Zimbabwe. However, in Customer Satisfaction Performance we listen to Voice of the
Customer; in fact it is one of the errors that providers of goods and services make: turning to
‘norms’ however undesired they are.
The process of Customer Satisfaction Performance analysis was however not in the
224
perfection of typically QFD-organisations. The Student Evaluation of Teaching Effectiveness
instrument and the instrument used in surveying stakeholders were designed from research
and collation of a number of similar instruments used elsewhere by other institutions. In real
QFD contexts and where the market is a special niche the ideal situation would be for
stakeholder representatives to work together in producing a system of metrics that derive
from concepts of the products and services they desire. In chapter 3 we lamented the
discrepancy among products and services, throughputs and metrics for control and
monitoring of quality. Students could check each question / prompt on the Student
Evaluation of Teaching Effectiveness form with some numerical figure and spaces for
commenting were availed. The data was not however computed into Weighted Average
Performance values. This shortcoming could mean that the CUTGBS is failing to segment
whatever products and services it could. Segmentation could improve Customer Satisfaction
Performance, cut on costs and improve institutional competitiveness. These were the same
reasons that staff mentioned as having motivated their decision to pursue a QFD-based
strategy. In a comparative case, two Business schools were muting more specialised Master’s
degree programmes than the more general ones.
A number of alumni and the 2013 students indicated that they had chosen the CUTGBS
M.Sc. programme in response to the CUTGBS marketing, word of mouth, and other factors,
including it having been the first response they received after their applications. Alumni and
students variably sited career progression, competitiveness and role modelling as reasons for
seeking postgraduate studies, with no specific preference for the CUTGBS being cited by
over 50% of the participants. Over 50% of participants felt that the CUTGBS was a
satisfactory provider of postgraduate studies. They however had numerous suggestions on
improvements, ranging from parking, catering and process alignment and interface mapping
to removal of what they felt were unnecessary procedures and requirements.
Directors in the ZimCHE felt that the CUTGBS was doing well and complying with the
regulator’s requirements [DR1, DR2, DR3]. An academic at the CUTGBS sited that the
ZimCHE had ranked the CUTGBS number 5, but did not understand the measures, standards
and procedures of the ranking [AC2]. A comparison of Customer Satisfaction Performance
225
with each stakeholder group generated a long list of qualities and standards of satisfaction.
This is evidenced in the preceding paragraphs. It would need a huge budget to move swiftly
to meet every requirement from every stakeholder. As resources are not mentioned in most of
the demands it makes sense to draw up a priority listing. In QFD environments organisations
would use the Kano model to assess and evaluate Customer Satisfaction Performance of
products and services. The CUTGBS was not deploying the Kano model in its deep technical
sense in Customer Satisfaction Performance analyses. This should be one of the causes of
discrepancy between planned quality and perceived quality. In traditional quality models
definition of products and services quality resides with the provider, in QFD it resides with
the customer.
In a nutshell the CUTGBS was fond of improving its Customer Satisfaction Performance and
was gathering customer feedback on its curriculum and other offerings. Students were
satisfied with the M.Sc. Programme as was the ZimCHE with the performance of the
CUTGBS. There was room for further exploitation of technology to process Voice of the
Customer data and escalate it into Product Planning Matrix so that products and services
align more strongly to Voice of the Customer.
Student participants agreed that it was too difficult to obtain information about any
programme in its completeness and that each business school ‘had some level of bluffing’
[IST1]. This bluffing and the rush for higher qualification were some of the push factors that
influenced enrolments in ‘anything that came first’. Franceschini (2002:146) expressed this
kind of discrepancy between marketing quality and offered quality mathematically as:
Gap = marketing quality – offered quality.
GAP = Qr - Qm
This discrepancy is attributable to the inadequacy in communications between individuals
and teams within the Department or sector and secondly to externally directed marketing
communications that emphasise what the organisation is already excellent at, while
downplaying its shortcomings. Jabnoun (2009) in reference to provision of MBA degree
programmes by for-profit institutions in Vietnam refers to the gap or discrepancy between
marketing quality and offered quality and attributes it to, among other factors, the stiff
competition for students. An alumnus said of the discrepancy:
“... yeah it is the ‘moment of shock’ you say so is this what the stakeholders they claim to
have consulted told them to teach us? ...” [FG2-ST4]
Paul Cooper (2007:19) talking of the same gap in UK higher education and elsewhere
attributes the discrepancy to quality uncertainty and informational asymmetry between
students and the provider institution. Akerlof (1970) suggests that some level of
misinformation is deliberate and directed at covering the shortcomings of ‘lemon
programme’ whilst wooing students. Literature generally refers to this practice as ‘market
signalling’, which has the risk of helping low-quality products and services displace high-
quality products and services from the market (see Cooper, 2007). Things as simple as
nomenclature of subjects can do a lot in marketing. As institutions do benchmarking they
227
need extra care, otherwise they benchmark ‘lemons’ that are over marketed in what has come
to be known as the ‘Enron Effect’. Complains may not actually be based on the programme
being of any inferior quality, but because of globalisation students may be comparing
programmes in a developing country with those in developed countries and largely
disregarding the local focus of the programme in their comparison.
In summary, the CUTGBS was improving the Competitive Satisfaction Performance of the
M.Sc. Programme and had in place a robust system of scanning the market and incorporating
trending issues into its management, products and services strategies. The analysis of
Competitive Satisfaction Performance was basically qualitative with the exploitation of
tables, charts and matrices not having fully cemented.
It is said in some literature that leadership, the number of books in the library, and computer
laboratory are some of the representations of quality. In this research a handful of students
did not think they would represent quality on the part of the CUTGBS because they already
had own laptops, could buy own e-books, and use e-journals to get better content. One
student jokingly said (s)he could teach the CUTGBS’s best Computer teacher ICT
(Information and Communication Technology), implying that she had good computer skills.
One student showed this researcher her (his) e-library with more than 50 e-books on his (her)
laptop and many students had personal laptops, iPads and tablets. Improving aspects that had
no impact on teaching and learning and which would be more accessible to students by their
own means would have no sales impact on the M.Sc. programme. In the case of a service like
228
the provision of higher education, target values may not be too easy to define, but still people
would have an understanding of what it is they aspire to do or have. Target values may mean
quality of facilities, competencies, process features, the intended learning outcomes, and
even mental frameworks. Participants variably indicated sentiments that there was still some
need to improve management and curriculum design in the teaching and learning aspects of
the M.Sc. programme. The same participants said they would recommend the programme to
third persons. The balance of subjects was appreciated by many students (see Appendix 3 for
structure of the M.Sc. programme).
In summary, I observed that selecting teaching objectives, teaching material and the
pedagogic approach has become more complex because in terms of skill and knowledge the
student may be quite near to or even surpassing the levels of the lecturer.
Teaching approach
Students said the structure of the teaching was good particularly that they received handouts
electronically for most subjects before they come for the sessions and that some lecturers
continued to send them more information after the sessions. However, students felt that they
were not maximising their gains from the M.Sc. programme because of the short duration of
the programme. Most students did not think that cutting the number of subjects was a
solution, but rather extending the duration of the M.Sc. programme from 18 months to two
years. There were plans by July to introduce e-modules and students were welcoming the
idea of becoming ubiquitous learners.
Resources
Academics felt that the locus of decisions that had to deal with procurement and use of
resources lay more with the mother university than with the CUTGBS. CUTGBS did not
actually worry about the use of finances, but about the speed of responding to their requests
for different procurements. The tension appeared to be one of policy and the time-based
nature of the CUTGBS where agility and a desire to be out front dominated most decision
processes in the CUTGBS. Resources are a critical factor in policy dispensation as they are in
strategy implementation. The implication is that there is need for a strong policy-strategy
229
alignment that should allow for appropriate resource utilisation. This reflects the point
suggested at Section 5.3 that to meet the challenges for quality and quality assurance in HEIs
it is critical to harmonise the formal design infrastructure and the institution’s emergent
designs.
Consultancy work
Fundraising through consultancy was highly encouraged in the Chinhoyi University of
Technology. Staff at the CUTGBS were welcoming the idea, but still wondering how much
time they would have for research, consultancy and teaching. The relation between
consultancy work and QFD is that QFD improves efficiency, resulting in huge financial
savings. These must be invested to raise even more capital. It also opens opportunities for
staff and students to have a brush with the industry. Temtime and Mmerrki (2010) observed
the need for staff to have in-depth understanding of the industry in order to improve their
effectiveness at teaching master’s-level courses that were industry related.
Conditions of service
In the QFD context, target values are not exclusively intimately directed at the products and
services. Target values both for academics, other staff and the CUTGBS as an entity were
being set. Academics felt that their remuneration package was satisfactory. The package
included cost to company facilities like transport, housing, and teaching allowance that was
rated per hour of classroom teaching. They were paid for every cost incurred while on
business outside of the university, sabbaticals and attendance of workshops and conferences.
There was a shared feeling that they needed to improve on team-working skills, particularly
when it comes to membership with others outside of the CUTGBS. The feeling that the
CUTGBS would work more effectively and efficiently with greater autonomy on a number
of issues was not repeated all too often. There were concerns that the current ‘institutional
culture’ (August 2013) was in ‘many ways incompatible with the ideal strategies of today’s
business schools that aimed to be relevant’.
231
Target values from the viewpoint of industry and alumni
I held focus group discussions and one-on-one discussions with CUTGBS alumni and two
established recruitment agents (see Chapter 4). I continued communications with each of the
focus group members after the first interviews. An analysis of the data collected showed 19
conspicuous points that alumni and personnel from industry confirmed they found really
important in the industry at almost any level of employment. The factors had to deal with
academic intelligence. This correlates with W.R. Scott (2008:37) who mentions the value of
a broad-based curriculum that builds manager skills of synthesis and evaluation. The second
category of skills falls within individual’s emotional intelligence, which covers self-
awareness; self-management; relationship management; and social management. The third
related to spiritual intelligence; an area that covers characteristics of maturity, compassion
and humility. These academic or curriculum related target values correlate to the
characteristic of ‘ideal master’s graduate’ discussed in this thesis at chapter 3.
In summative terms, there are challenges with precision in the defining and finding the best
fit between the HEIs’ ability to know and deliver what exactly matters for expected quality.
If Target Values could be defined with precision then curriculum goals would be easier to set
and Improvement Ratios even easier to operationalise with the consequent result of taught
curriculum (offered quality) approximating more and more to expected quality.
In this model there will arise six classes of coherent strategies from the self-assessment
sessions:
Strength-weakness strategies that use organisational strength to overcome its own
weaknesses.
233
Strength-challenges strategies that use the organisational strength to confront its
challenges.
Strength-opportunities strategies that use organisational strengths to exploit opportunities
and optimise organisation’s performance.
Opportunities-challenges strategies that use the organisational opportunities to confront
its challenges.
Opportunities-weakness strategies that use the organisational opportunities to confront its
weaknesses.
Weaknesses-threat strategies that aim to relieve the organisation of any facet on it that
attracts the challenges and those that create or depend on the facet that the organisation is
weak on.
Secondly the CUTGBS was focusing on correlation of offered quality with marketed quality
through marketing for Six Sigma (MFSS) by improving teamwork quality (TWQ), intra-
CUTGBS communications, and communications between the CUTGBS and other sectors,
and being up front. The CUTGBS was focusing training, coaching and mentoring on learning
needs in the hope of improving customer satisfaction, doing some interface mapping,
focalising on local leadership and the adoption of technology to reduce failure modes
(TFSS). All these efforts were directed at correlating planned quality with offered quality.
The CUTGBS was heavy on Voice of the Customer but quite light on its subjection to the
design of the four Six Sigma roadmaps I explained in Chapter 3. With more open, teamwork
based organisational structure in which more decisional and strategic power resides with
academics there were huge potentials for approximating expected quality to hypothesised
quality.
In a nutshell, the CUTGBS was actively trying to align its goals infrastructure with its
objectives network, right from policy-to-strategy correlation through to perceived quality and
offered quality and the strategy to the aspirations and ambitions of its staff.
In a summative way, the CUTGBS was resolute on that the quality they sought was based on
‘fitness for purpose’ and that transforming the knowledge base, attitudinal make-up,
understanding, belief system and behaviour was their key means of achieving fitness for
purpose.
236
discusses modes of institutionalisation and springboarding from that analysis I propose the
above framework in which I indicate the mode of adoption taken by the CUTGBS. The
CUTGBS adopted every stage of the QFD model as shown in this section. There was
however elements of selectivity in the deployment or use of some tools and techniques (see
Appendix 5 and Appendix 6) as was with respect to some conceptions. Figure 5.6 represent
this analysis.
5.6 The intensity of use of QFD tools and techniques and the implications on the level of
adoption of QFD.
5.6.1 Introduction
The purpose of this discussion is to examine to what degree staff deployed the various QFD
tools and techniques with the intention of embedding them into the infrastructure of the
structures, procedures, processes and other activities of the CUTGBS. Doing so answers the
research question: How effective was the implementation of the QFD tools in the M.Sc.
Programme? The analysis is based on the understanding that adoption of a strategy
framework happens on a continuum from sloganeering through shallow to deep embedment
of the model’s tools and techniques in the modus operandi of the adopting organisation. The
mix-and-match of tools may differ according to contexts which include perceptions, skills
and the configuration of the organisation’s factor dependency field or the organisation’s
emergent design.
These three propositions advocate for a new perspective and understanding of excellence,
value for money and consistency as means to improved quality performance. In essence,
QFD challenges the Harvey and Green (1993) definitions of quality. Studying the transcripts
one realises that CUTGBS staff frequently made reference to their use of the techniques in
Appendix 5. The purpose of the focus on the tools was on their appropriate use, depth of their
use and frequency with which each tool was used. I found a number of vectors along which
the strategic tools were being used:
information gathering – interviews and focus groups,
environmental scanning – environmental scanning,
competitive analysis – benchmarking, multiple scenarios and product life cycle analysis,
futuristic projections – forecasting, future studies, trend analysis and sustainable growth
model,
knowledge building – Delphi Method, Dialectical Inquiry, Nominal Group Technique,
strategy formulation – Critical Success Factors, Experience Curves, Market
Opportunity Analysis, Metagame Analysis, Multiple Scenarios, Product Impact Market
Studies Analysis, Situational Analysis, Strategic Gap Analysis, Strategic Issue Analysis
I do not make claim that this classification is watertight. It is my own distribution of the
techniques based on how I interpreted my participants’ narrations and my perfunctory
comparison with literature. Some of the tools could fall into more than one category but I
have decided to ignore the weaker in favour of the one with a stronger link.
238
Frequently used tools
The tools most used relate to strategy formulation or planning. This reflects that the adoption
of QFD had opened the scope of activities for the CUTGBS team and that there was a lot of
rethinking / reorientation in a number of areas as well as the shift of certain roles from the
mother university to the CUTGBS and these roles drag with them the need for use of those
techniques. Shah and Nair (2014) refer to academics as disinterested in strategic planning.
The situation was quite different with the CUTGBS staff. Also among the frequently used
were those tools relating to futuristic projections and competitive analysis. The set of such
tools speak of an aggressive institution looking for a secure niche and a robust strategic
balance. AC-1 said it was regrettable if any of the M.Sc. Programme lecturers was ‘sneaking’
MBA stuff into the M.Sc. Programme. This reflected the CUTGBS’s concern for a market
share and a competitive position and the CUTGBS was working hard to meet its goal of ‘20
% increase in market share’, ‘profitability’ and ‘survival’. The CUTGBS was very concerned
about its future and its competitive positioning. I find this high interest logical if we look at it
in terms of the initial motivation for the adoption of QFD expressed in Section 5.4 above.
Futuristic or forward thinking is a basic engagement of strategists in times of new product
designs and during setting up of systems in competitive environments.
The CUTGBS core team is composed of Business, Management and Marketing lecturers
who appear ready to show optimal distinctiveness by a record of success. This is reflected in
the quotations at Section 5.4.2. The CUTGBS was heavy in the use of tools relating to data
gathering than to escalation of the data into operational strategies. This should not however
be construed to mean that this is a permanent feature of the CUTGBS. Data gathering, data
processing or results analysis can feature prominently at different stages of project
implementation and presumably institutionalisation of QFD. Most of the tools were fairly
241
used. The intensity of use of tools and techniques depends on project size and the level at
which QFD is being implemented. With a small project like a M.Sc. Programme the use of
some tools could cover for the non-use of other tools.
In summary, the CUTGBS was using most tools used in QFD and strategic planning with
variable intensity and profundity that matched the importance the CUTGBS was attaching to
different stages of QFD. There was evident need to understand its customers, to plan
products that aligned to customer needs and to generating market intelligence. In the next
section the thesis discusses how the CUTGBS responded to results of both internal and
external quality management.
5.7 Staff perception on the implementation and institutionalisation of the QFD model for
the purposes of quality assurance in the M.Sc. Programme.
5.7.1 Introduction
The purpose of this section is to discuss staff perception of the management and use of QFD
to ensure quality in the M.Sc. Programme in response to the research question: ‘‘What were
the perceptions of staff to both internal and external quality assurance interventions?’’
The feeling that QFD could help in internal quality assurance was widely shared within the
CUTGBS. Most exhorted in QFD model was its provision of a roadmap of doing Voice of
the Customer and linking it to processes that guaranteed Customer Satisfaction Performance.
242
5.7.2 Response of the CUTGBS to internal quality assurance
The CUTGBS had adopted the QFD stages as part of its internal quality assurance
infrastructure. By so doing it had tried to guarantee that every stage of products and services
generation was linking to Customer Satisfaction Performance, a point I highlighted in
Chapter 3. Staff were required to have detailed lesson plans, mark-lists and a host of other
stuff which some felt uncomfortable with. One staff felt these requirements were reminiscent
of those for primary school teachers. However, some staff felt that the requirements of
detailed lesson plans, remediation and assessments were a good thing as that would show
how the teaching and learning was going on. Self-assessments and audits were being
encouraged. There was evidence that the CUTGBS was drifting towards taking fuller
initiative in quality assurance work.
Staff emphasised that it was basing internal quality assurance on: collecting Voice of the
Customer and responding to it [AC-4]; restructuring the CUTGBS to achieve a strong
quality-focused team [SM-1]; improving curriculum focus, content and processes [AC-1]; to
align training to customer satisfaction [SM-5]; to build the strategic capability required for
sustained quality assurance work and ultimately making a QFD approach the CUTGBS
instrument of quality [SM-1]. The CUTGBS had also taken the ZimCHE instrument and
escalated it to policy and a guide for internal quality assurance. In this light what is discussed
below fits both internal quality regulation as well as external quality assurance requirements.
Some of the points raised here may sound pedestrian. Those familiar with the discipline of
QFD should understand that QFD is about taking an organisation from the state of some
currently undesired status to a new and better one. It is therefore possible that a non-QFD
organisation can be having superior performance in one or two aspects when compared with
a QFD-based organisation at some moments in their development.
Despite these setbacks the arrangements of seats allowed for clear vision of the screen,
flipchart stands, and the whiteboard. There were safety features like emergency escape
routes, fire-horses, sand buckets, and fire extinguishers. The CUTGBS had reconfigured a
majority of the infrastructure to align with needs of master’s level students and also with
technology, evident use of Technology for Six Sigma (Voice of Employee + Voice of
Market).
245
Minimum number of students
ZimCHE ACCR 2A form says on page 2 that ‘Where a programme is so popular that a large
number of students have to be enrolled, the number of lecturers should be increased
proportionately’, but one academic says that the number of students was ‘... just too large,
yaa you see when numbers go around 50, 100 that’s too much for a subject and a single
lecturer’ [AC4] and another said ‘... sometimes the number factor becomes a burden on
quality initiatives, particularly ...’ [AC5]. AC3 admitted that it was better for those taking
electives as they would have smaller classes. Despite the mediatory effect of technology that
could allow class sizes to be larger than 10 students the economic concerns tended to
override the quality concerns and overwhelm the regulations on minimum class size. The
situation was not different for computer classes which were pegged at a computer-student
ratio of 1:5. However some students were bringing into classes their own laptops and tablets.
Students desired more to be taught how to run certain programmes on their laptops and
tablets than to use university computers. An academic had observed that use of the computer
laboratory time had changed with student requirements tending towards request for
installation of some programmes and small issues with their gadgets. Previously it was about
teaching the student how to open the computer and learn how it works.
246
he had never used the common portal [AC5] despite his / her subscription to the idea of a
common portal. The Academy of Teaching was running induction-cum-training sessions in
which lecturers were among other things trained on assessment, evaluation and examination
techniques. A senior Training manager said that the ‘teach-the-professor’ project was borne
out of a realisation that there could be some misaligning of the curriculum from the synopsis
of the M.Sc. Programme to the examination question and the marking schedule. An academic
in the CUTGBS said they wanted to see the idealised M.Sc. Programme graduate in the
question paper and on the moderated script [AC1].
Many felt that fear of exposure was a key deterrent factor. A senior university manager
[SM6] explained that they operated an examination paper bank and no one would ever know
which paper would be raffled out for the particular examination. ‘‘When a few minutes
before an examination starts we conduct a raffle of more than ten papers who could have ever
dreamt which paper was likely to come?’’ Lecturers were required to do detailed lesson
plans with objectives, methods and assessments. Internationally various types of unethical
behaviours have haunted higher education including sex for marks.
5.8 Management’s response to the results of the application of the QFD model
The QFD model was adopted within the strategic desires of the Chinhoyi University of
Technology to proffer quality education to the nation. Being a new adoption it raised anxiety
in management because whatever the results of its implementation, would have far-fetched
implications across the university. This section discusses the research question: How did
management respond to the results of the application of the QFD model? The idea one would
get from narratives of participants was that for the CUTGBS, QFD was an approach to
management that ensures that every process, structure, activity, tool or technique and their
interfaces were working optimally to render highest levels of quality products and services.
In the structure of QFD shown earlier in this chapter is an assessment-evaluation loop for
each stage of the QFD. In essence the CUTGBS was running intra-stage assessments and
evaluations. These would also be run posterior to every stage and both reactive and proactive
action was being taken. Running the Voice of the Customer had helped the CUTGBS realise
unexploited niches and opportunities to which it was responding by launching in 2015 a
master’s degree programme in Business Intelligence. The response of management was to
support the launch of the M.Sc. in Business Intelligence Programme. Academics in the core
QFD team explained that they had come up with the idea of the Master’s in Business
Intelligence from their analysis of Voice of the Customer [AC1]. The industry and students
were said to have suggested that the CUTGBS consider some content areas that were not in
the M.Sc. Programme curriculum and others had requested that greater emphasis be placed
on some areas than others [AC3]. The CUTGBS was also launching short courses in other
discipline areas like Project Management.
248
Implementation of QFD had shown that some of the training offered at the Academy of
Teaching was not meeting some of the needs and requirements for effective implementation
of QFD. The CUTGBS had began to augment university-wide training sessions with their
own upskilling projects designed to improve skills in areas of CUTGBS special needs.
Technical competences were being emphasised, including the use of software in running
different programmes. All these projects had been approved by the university management
from the level of Academic Board, Faculty, through to University Council. We take the
assumption that the approvals were a sign of positive response to what QFD was achieving
for the CUTGBS. An academic within the CUTGBS was offering ‘sit-by-Nelly’ training to
peers in statistical and mathematical areas including how to design some matrices.
The CUTGBS was eventuating into a high-tech institution with most decisions being data-
based. University-wide training sessions had special focus on teaching, curriculum design,
and other technical-administrative issues. Both the Chinhoyi University of Technology and
the CUTGBS had learnt that there was need for more autonomy on the part of the CUTGBS.
I have referred to this as enriched autonomy because both sides were realising the context-
driven need for that amount of shift in decision-making loci. The CUTGBS was increasingly
becoming the locus of many important decisions. The increasing support and commitment of
management to CUTGBS projects launched under the auspices of a QFD approach could
mean that the model was helping in generating positive response from management.
Organisational structural designs and managerial actions have an undeniable impact on team
processes including those resulting in generation of individual and organisational identity and
in decision-making processes in the team. The current granting of decisional power to the
CUTGBS was giving it the strategic agility that goes with competitive markets. Most
requirements from the ZimCHE had been escalated to policy and standards of work in the
CUTGBS.
The CUTGBS started off with a small number of PhD holders and felt the need to upgrade
the quality of its staff by sponsoring oversees PhD studies. University management took on
249
the idea and by 2013 eight staff were being sponsored for international PhD studies at
university full cost.
The university was also supporting any CUTGBS staff including with paid leave for any
study endorsed by the CUTGBS management. The implementation of QFD in the CUTGBS
had influenced management to give a status almost like a pilot project with little regulation
and extended support. These responses to the CUTGBS were being found encouraging by
CUTGBS even though they felt these were but rolling out slowly in some cases.
“If training everybody in anything is what quality is, that is there. But if it is what I
understand it, collating activities across the institution and subordinate them to a quality
objective that one is not there.” [AC1]
AC1 was observing some confusion about what a ‘learning organisation’ is and
‘organisational learning’ should be.
The Academy of Teaching was actually running induction-cum-training that took ‘serious
note of the requirement to see that lecturers have appropriate teaching skills’. Lecturers felt
the training offered was important and a worthwhile effort. The CUTGBS was taking part in
such training and in one interview academics AC-1, AC-5 and SM-1 explained that
management was thinking of the CUTGBS be a source and designer of some training
particularly on strategy focusing and quality management. The feeling was that the CUTGBS
was scoring well on the Balanced ScoreCard and should be a good instructor on how it is
done.
The use of Voice of the Customer had shown a number of necessitated changes, like in the
naming of subjects and restructuring of the M.Sc. Programme. A request was forwarded to
the University Senate and was quickly sanctioned. AC-3 in reference to this says that the
CUTGBS has become known for following Voice of the Customer and that the management
appreciated the importance of their proposal because they knew it reflected what matters to
250
us, to the university and to the idea of Customer Satisfaction Performance. The Research
Methodology subject had been moved from semester 3 to semester 1 to enable students to
begin framing their research projects immediately. Academics variably explained this in
terms of Target Value analysis and Product Planning Matricing, both which are
manifestations of QFD’s Voice of the Customer.
The CUTGBS had muted a plan of a ‘state of the art’ campus which had already been
endorsed by the management. Staff in the CUTGBS felt QFD had thought them a ‘kind of
way of thinking’ and ‘getting things done’ through creating own opportunities [AC-4].
In other ways management was responding to the results of the application of QFD in the
CUTGBS by supporting, providing resources and taking a detached engagement approach.
Trust with the ZimCHE was said to be strengthening and a Director with the ZimCHE said
that the Chinhoyi University of Technology and the CUTGBS were very responsive
institutions.
5.9 Conclusions
In this chapter the nature of QFD was discussed, plugging perspectives of scholars,
practitioners and the CUTGBS staff into each other. QFD offers a philosophy, methodology
and the toolkit for assessing and evaluating quality in the M.Sc. Programme and the
CUTGBS. A new epistemology of quality, with new courage to initiate unfamiliar ways of
quality assurance is long overdue. In the new perspective shown by this research the
traditional definitions of quality are challenged. The possibility of a concept of quality
defined in terms of a conscious purpose rather than a constituent-based one is advocated. The
depth a model penetrates and cements in an adopting organisation can best be measured by
the intensity with which the key facets of the model are used. The CUTGBS was responding
to QFD by adopting of its stages and some of its tools. Creative and context based adoption
are a better strategy of model use because that gives the adopter room to self assess, to built
the necessary resources and strategic capabilities at each stage without choking self. The
Chinhoyi University of Technology was learning from CUTGBS and were providing
251
solutions to every shortcoming exposed at each key QFD stage. QFD flags shortfalls and
show how these can be remedied.
252
CHAPTER 6 – DISCUSSION, RECOMMENDATIONS AND IMPLICATIONS OF THE
STUDY
6.1 Introduction
This chapter discusses the research findings and proposes recommendations and implications
for various stakeholders. It should be borne in mind that this discussion on QFD, unlike most
discussions of QFD in most research like those cited in chapter 2, is about all the stages of
QFD because the model was adopted in its wholeness.
253
The interwovenness of these elements will influence the texture and flow of the discussions
in the sections that follow. The research contributed to our appreciation of quality and quality
assurance and to a growing but still limited understanding of the adoption and
implementation of New Public Management (business) models by examining how the
CUTGBS adopted and diffused the various stages and tools of QFD and its implications on
the quality of the M.Sc. Programme. Research in education has developed rather slowly, is
widely diverse and is limited in scope (Newton, 2000:153; Ehlers, 2009:344; Haggis, 2009).
This has resulted in an archipelago of miniature advices and recommendations that
practitioners find difficult to see how each plugs in with the rest (Ramirez, 2013). In the
CUTGBS most staff combined industry and academic backgrounds in areas where numerous
aspects of QFD and managerialism are commonplace. This seems to have contributed to the
CUTGBS’s confidence and ease with QFD. I wonder whether that amount of ease could
characterise people without prior expose to some elements of QFD and managerialism. QFD
uses the Theory of Constraint to incessantly identify constraints, resolve them and handle the
process in a more holistic and synergistic manner.
Contributors to final validation appraised the feasibility of the QFD model where CR-2 and
CR-6 specifically stressing the need for providing QFD-based training on an aggregate basis
as well as on a role-based approach. These comments tie in with the CUTGBS’s efforts to
augment university-wide training sessions with own learning-needs through focused in-house
training. It also ties in with recommendations abounding on QFD / Six Sigma blogs. In
literature Franceschini (2002), Ficalora and Cohen (2009), among others, stress the need to
continually raising capacity of staff right from White / Yellow Belt level to Master Black
Belt levels. Such in-house training can be locally certificated which the mother university
was already doing in response to Voice of the Employee. From this background I discuss
what QFD meant for the CUTGBS, why it was adopted and how it was being implemented.
Contributions from the validation study stressed the most ignored part of model adoptions:
the impact of the model on management and the amount of change that the model will
require on the part of management philosophy and practices. CR-5 said that QFD was
working in the CUTGBS because the CUTGBS management had surrendered to the model,
accepting it to give them a new way of understanding quality, management and processes.
CR-2 stressed that no NPM model will sustain in traditional bureaucratic management.
This was a case of a context (problem) looking for a model. The inverse is where
management chooses a model and dictates its adoption, like medicine looking for a patient or
a solution looking for a problem. A clear understanding of the context should be a better
point from which to share visions of the future and what models can best connect the
organisation with its context and the shared vision. Good models must cater for the human
aspects, resources aspects and relational aspects of an organisation. QFD was adopted in a
rather creative way. In chapter 2 I discussed most adoptions of QFD but these involved
mainly the Voice of the Customer component of QFD and the adoption of Product Planning
Matrix. There was also positive response to the way QFD was being institutionalised.
Contributions from the validation study were that Voice of the Customer was undoubtedly a
critical component of newer strategies aimed at improving relevance of curricular to industry,
society and students. CR-1, CR-3 and CR-6 particularly appreciated the efforts in linking the
M.Sc. Programme to Voice of the Customer. They were joined by the other contributors in
that at the national level the other pillars to quality: Interaction between the ZimCHE and
Students; between Industry and Students themselves was developing rather slowly. In more
general comments CR-4 felt that addiction to traditional ways of doing things was delaying
the inclusion of technology and an abundance of software in processing of Voice of the
Customer and in their use in curriculum development efforts.
.
257
Figure 6.2: The relational nature of context, model, strategy, actions and outcomes.
258
Staff’s positive response to the QFD model was further shown by the commitment shown by
CUTGBS to implementing every stage and most tools of the QFD model. There was also
positive response to the way QFD was being institutionalised. This could be drawn from how
staff explained their teamworking through all aspects of the QFD methodology and how they
expressed a feeling of strategic bundling. This positive atmosphere could have been
generated by the active inclusion of staff members right from the selection of the model and
the increasing strength of the core team and the local leadership approach. The increasing
shift towards a local leadership approach was affording the CUTGBS team members more
clout and influence as strategist (see chapter 2). For some, QFD was becoming the vehicle of
what Senge et al (2005:76-77) call ‘presencing’ – the state of becoming totally present to
what is happening inside oneself, around oneself and what is emerging through oneself.
Figure 6.3: QFD influencing the way strategy planning, change-project management,
implementation and risk management were done
259
Models translate into strategies the moment their elements begin to be implemented. QFD
was being worked to instil a culture of kaizen or continuous improvement by influencing key
organisational processes as shown in the figure below. The three-prong approach to
improvement covered human skills, teamness intelligence and improvement of impersonal
infrastructure. QFD was becoming a game changer by improving clarity and focus in
strategic planning and in aligning and integrating processes and goals in ways that made the
strategies work better. The 13 stages of QFD were treated as mini-projects giving the
CUTGBS a strong sense of what needed be done, why and how. A special form of systemic
and systematic thinking was evident. Apparently the CUTGBS had recognised and were
exploiting the self-assessing nature of the QFD model. QFD was facilitating the management
of model-based risks and of extra-model constrains. The CUTGBS had placed QFD at the
heart of its life. A model is as handsome as it does, just as the worth and value of a vision is
measured by what it gets done on the ground. Assessing staff response to a model should
have implications on the design of structure-structure, structure-function and function-
function relationships as on conditions of work and remuneration so that conditions for
catalysing and enabling are reinforced. The quality of a response has a knock-on effect on the
intensity of use of model tools and their diffusion rate.
Reflections during the validation sessions were that as a new model to a majority, there was
need to proffer training and to take time and patience with those expected to enact the model
and to understand it. CR-2 pointed out that every stage of the QFD was possible and
necessary and there is need for showing members why every tool is important, how it is
important and how to learn how to work with it.
McMurty (1991), and Parker and Jary (1995) had earlier classified academics as ‘sinking’
under the supremacy of management, something equally suggested in Jnanesh and
Hebbars’(2008:2) model of correlation matrix. Newton (2001) describe two perceptions: a
feeling of intrusion, inspection and bureaucracy and the other of conformist behaviour.
Nothing of the categorisation of McMurty nor of Newton or Towler’s sinking, was evident in
the CUTGBS. A common perception of quality echoed among students, academics and QFD-
based literature is fitness for purpose. But whose purpose would it be if not the one defined
by the four voices: student; academic, market (industry), business (quality regulator)?
Building from transcripts and the development of categories Harvey’s constructs of value for
money, consistency and exceptionality adopted a new meaning in the context of QFD. Figure
261
6.4 below link them to the Six Sigma roadmaps. If the student-industry is the primary
customer of education then improving student knowledge, attitudes, skillset, understanding,
behaviour and belief system should be the primary purpose of education. The essence of such
improvement consists of transforming the student from current to new and superior levels
through a transformative instructional process. The vectors of change that lead to such
transformation must embed four issues. One is Design for Six Sigma (DFSS) which is
directed at reducing and eliminating constraints (structures, relationships and mentalities)
that make the organisation dysfunctional. The second being Six Sigma process design
(SSPD), which is directed at reducing or eliminating processes that waste resources and make
the organisation less effective. Thirdly would be Marketing for Six Sigma (MFSS), which is
directed at reducing or eliminating relations and external competences that don’t improve
customer satisfaction. The forth being Technology for Six Sigma (TFSS), which is directed
at use and adoption of technology that enhance Customer Satisfaction Performance.
Figure 6.4: A holistic construct of quality as synthesised from literature and transcripts
262
Achievement of DFSS saves resources thus allowing for their deployment into stages and
processes that increase net value to student. This deliverable is what both Harvey and QFD
call ‘value for money’, definitely not a definition of quality but a pre-condition for quality.
With increased effectiveness and efficiency value for money increases because costs are
reduced thus the felt value-for-money effect goes up. Achieving Six Sigma Process Design
implies that organisational processes are subordinated to and consistent with Customer
Satisfaction Performance, definitely not a definition of quality as advocated by Harvey.
Achieving Technology for Six Sigma means the organisation is adopting exceptional
(excellent) technologies that have a heighten impact on Customer Satisfaction Performance.
Thus exceptional or excellence does not define quality but is a pre-condition for superior
quality.
The CUTGBS’s House of Quality shown above is holistic in that it proffers a roadmap for
front-end incorporation of voices of professional organisations and quality assurance
agencies rather than the separate treatment they receive in traditional models. Apparently
there were issues with working each of the four roadmaps. These included minor resistance
to structural reconfiguration, adoption of newer processes and revoking deeply seated
processes, redefining metrics for market success and negotiating power over definition and
verification of technology for acquisition. The positive response to QFD and its becoming the
overarching statement of an epistemology or philosophy as well as methodology for quality
assurance generated an innovative approach of (re)looking at quality. However it raised
questions of governance, model diffusion, strategy and purpose from university management
on the CUTGBS. What should be the optimum amount of autonomy to grant out? How
would the deliverables of QFD influence other sectors of the university? And how will ideas
about strategy and purpose impact thinking, behaviour and procedures across the university?
Acceptance of QFD as a driver of change in terms of a quality paradigm means we should
ready ourselves to rethink who ought to hold the prerogative of defining quality, to rethink
about matching models of quality with appropriate measures, standards and metrics, and to
validate the model of quality proposed in the figure above. A deeper analysis of literature
actually show a huge convergence in ‘fitness for purpose’ as the fundamental construct of
quality: students want to be fit for (self)employment; industry want graduates who are fit to
263
be employable; society want people who are fit to propagate goodness. Lomas (2002) say
that HEIs voted ‘fitness for purpose’ followed by transformation as most closely matching
their definitions of quality.
Validation results suggested what most participants in the main study felt – that staff
response and perception of QFD was positive. CR-3 and CR-1 suggested that model
sustenance is a complex function of how those protagonising or ‘case-owning’ the model
make it enjoyable and worth-while to be involved with the model.
Management was giving necessary support in terms of advice, resources and remodelling the
accounting requirements. The CUTGBS staff were apparently trending toward corporate
federalism in which they wanted greater autonomy and treated other sectors of the university
as customers or markets. The university management was apparently satisfied with the
performance of QFD and were approving (2013) the launch of new master’s degree
programmes and short-courses. QFD had facilitated in drawing up a clear and convincing
Project Plan and business case for each of the programmes. I presented the following
responses of management to application of QFD in the CUTGBS: the granting of greater
decisional powers and more areas of responsibility; allowing the CUTGBS to run its own in-
house training rather than giving over such responsibilities to the Academy of Teaching;
treating the CUTGBS as a pilot in the implementation of QFD.
264
Contributors generally felt that giving lots of autonomy was a ‘win more – win more’ type of
strategy to both the mother university and the CUTGBS. CR-4 specifically pointed out that
policy issues in traditional universities are incompatible with new public management
models like QFD. In this light giving the CUTGBS a ‘pilot’ status would free it of the
constraints of such inhibitive policy regimes.
Voice of Customer
Voice of Customer is the entry point into QFD’s House of Quality and recommendation is
made as to the value of collecting data about all stakeholders, processing it and transferring it
into a database for factoring in future decisions. A better understanding of Customer
267
Satisfaction Performance and customer requirements should help in approximating offered
quality to expected quality. However this is a function of the rigour with which the Voice of
the Customer is processed and escalated into policy and governance.
Regulatory requirements
The study observed that the CUTGBS complies with the basic technical, normative and
regulative demands from the ZimCHE. It is good business to abide by regulations within the
area of business as this reduces the frustrations that follow penalties and disputes with the
regulator.
268
Figure 6.5: The different demands and their influence on graduate attributes as
synthesised from literature: the context that should inform course design
The important determiners of net academic, technical and social competences of the master’s
degree graduate are the results from Research and Development, quality assurance agencies,
the market and business. Extending this idea beyond a programme challenges this perception
as inadequate because it reflects a new or improved methodology but without a
commensurate philosophy. During the validation sessions, participants felt that a philosophy
/ epistemology or mental model of management could fit in this scheme as shown below. The
idea here is that HEIs should define quality in terms of fitness for purpose from whence they
define the characteristics of the ideal programme graduate. An understanding of the ideal
programme graduate should inform decisions on the tools of instruction and the methodology
of deploying quality at the programme and lesson level. But this alignment and integration
can only arise from an enduring philosophy or epistemology of quality that should be present
in academics, management and students alike. Figure 6.6 show this relationship.
269
FITNESS FOR IDEAL MASTER'S TOOLS AND METHODOLOGY
PURPOSE GRADUATE EDUCATIONAL OF FUNCTION
•Needs and wants of •Vectors of INFRASTRUCTURE DEPLOYMENT
society; industry; development in •Research and •Designing for
individual student knowledge; attitudes; development vectors; reducing failure
•The nation's skills skills; understanding; PHILOSOPHY and
approriate modes;
priority framework behaviour and belief technology use; •Consistency of Voices
EPISTEMOLOGY
system process of Stakeholders- OF QUALITY
•Curriculum scope management; interfaces-strategic
resources categories and
optimisation etc cartography
•Consistency of input-
throughput-ideal
candidate
Figure 6.6: The relationship between purpose of education and philosophy of achieving that purpose
270
Customer Satisfaction Performance
The running of satisfaction surveys is a welcome activity that should be sustained, however the
tools need some attention so that they allow more space for the stakeholder to vocalise more of
its own views. A 360-degree understanding of customer satisfaction is of strategic value because
it helps in scoping and directing future strategies.
Generally the CUTGBS is apparently doing well with the implementation of QFD and the effort
is worth commending. Strategies take time to settle and their institutionalisation is not without
its hardships. It is therefore important for the CUTGBS to learn ‘tricks’ of strategy
implementation and continuously collaborate with those willing to make QFD work.
By deeply analysing the nature of QFD, showing its duality the research has proposed a
philosophical and methodological frameworks and related tools and techniques for
implementing profound change in our thinking and doing of quality and quality assurance in
higher education. A new epistemology for instituting quality must advocate for profound change
in our mental model, in the building of strategic capabilities for relating processes differently
and behaving differently.
By drawing into the analysis the four Six Sigma roadmaps, the study has shown the huge
feasibility and desirability of the different constituents to adopt new mental models on who has
the prerogative of defining quality and how networks can work for the good of quality. With a
new epistemology of quality, people will draw into the fold of higher education the disciplines
of shared vision, team learning, presencing in the new, and collaborate for superior quality
performance.
271
Another contribution of the study is its verification of the feasibility of the whole QFD model in
education. Most research on the application of QFD in education has been with its part-
components not about the implementation of QFD in its totality. From the main discussion in
Chapter 5 and the validation study the implementation of QFD’s stages (House of Quality),
whilst feasible, it requires patience, and continual focus on capacity building and risk
assessments. Therefore to attain and sustain the momentum for quality performance it needs
patience, going slow and ensuring continual building of strategic capacity for more change and
stronger focus on quality.
The research has shown that traditional models for quality assessment and quality assurance are
monolithic and zero on small aspects of the education infrastructure. QFD, apart from
encompassing the many different models, it ensures that they are contextualised and redefined
in relation to the needs and wants of the customer. Each of the models orients people relative to
a predetermined aspect of quality and quality assurance not relative to the broader higher
education landscape. But QFD proffers a way of effectively organising them in my proposed
representation of the House of Quality for the CUTGBS’s M.Sc. Programme (Figure 6.4).
The study has shown that a focus on improving quality without subordinating strategic planning,
change-project management, strategies implementation and management of strategic, systems,
operational and unforeseeable risks to a model of Customer Satisfaction Performance fails. I
argue that QFD sharpens and brings in the discipline of systems thinking to quality and quality
assurance work. In relation to this alignment and integrative work, QFD is a game changer.
Traditionally organisations adopt strategies then do an environmental analysis. The study argues
that a context analysis should come first so that strategic issues are highlighted before a best-fit
model is adopted. Strategies that emerge from the datum of concerns should survive their brush
with the roughness of operational reality better than do those that are chosen and retrofitted into
the contexts.
By running the loop ‘analysis of results thus far’ the QFD model, more than any other strategy,
ensures the process of a deep-going self analysis and an external assessment and evaluation of
272
its impact on quality assurance. Education has not had such a powerful model that run its own
internal assessment. This aspect is further reinforced by running:
DFSS (Voice of the Customer + Voice of Business (QAAs; professional bodies, etc)
SSPD (Voice of the Customer + Voice of Employee)
TFSS (Voice of Employee + Voice of Market (Research; trend setters, etc))
MFSS (Voice of Market + Voice of Business)
Gathering the different voices and combining them then escalating them into the quality-seeking
strategies of the organisation removes the hegemony of any single stakeholder in defining what
quality should be produced.
The other contributory value of the research is the lateral thinking; finding equivalent language
terms for the very technical jargon of traditional QFD. Some stages of the QFD process are very
technical and cannot be directly transferred into education for instance reverse engineering of
competitor products. But organisations can study how they became produced and how they link
with market requirements. Whilst the value of QFD cannot be denied, the need for its
contextualisation is as important and may involve skipping, modifying, or surrogating some of
its components.
The other insight is that the research has defined profit not necessarily in terms of financial
gains. The study has looked at ‘profit’ in education as the ultimate gain expressed in
Improvement Ratio and in Absolute Importance as expressions of the growth in quality of
products and services rendered by HEIs. Some people erroneously assign QFD to for-profit
institutions where the profit is monetary. Efficiency, effectiveness and better service delivery
are equally prime targets for QFD.
The research’s other contribution has been the argument that for higher education institutions
(HEIs) wishing to adopt QFD, a forensic self-assessment should be able to show them the
resources and infrastructures already at hand. It’s much about inwardly looking at how to define
and structure the way one wants the issues of quality performance done. Most model or strategy
adoptions start with the model and then do a context analysis. This study argues that models
273
should emerge from a context analysis: start with context analysis, chose and adopt a best-fit
model, use the model to design and diffuse strategies, carry out strategic actions and monitor for
the desired results. The flow diagram below summarises this reasoning.
Figure 6.7: An approach for linking model to context, strategy and desired outcomes
The Validation Study showed that it is feasible to apply QFD as a strategy management
framework that can serve for assessing and evaluating quality assurance in a master’s degree
programme. Another achievement of the study is the advocacy of the Validation Study as a
strategy or way that examines the business meaning and implications of the results of a research
project. This helps shake and expand the validity of results. I feel this is needful for case studies.
A validation by experienced insightful people augmented with quick-study people drawn from
the peripheries of the decision-making and leadership cycles should improve the strength of case
studies and their generic value of research mobilisation.
The study has shown that the wish for a common quality cannot be sustained in the face of an
ever-increasing number of job-types, of careers, and diverse students. Quality should therefore
be defined by those who want it. Such a perception of quality is what higher education
institutions should study (context analysis), then decide the corresponding Quality model that
would suit the requirements of the customer (model’s emergence). When a model has emerged
from the analyses of the datums, the strategies for aligning the goal infrastructure and objectives
network can be drawn (strategy formulation) to guide activities (strategy implementation) so
that the outcome (offered quality) approximates as much as possible to expected and
hypothesised quality. This is the epistemology of doing quality in organisations that this study
274
highlights. We should begin by specifying what Voice of the Customer says then we design-
down the curriculum, the course outline and the management systems and processes. I propose
the path should follow this route:
Figure 6.8: ‘Designing down’ from ideal graduate characteristics to the course
This approach of designing down from the wants and needs of those attending, sponsoring,
claiming a responsibility to assure quality was validated as a way of ensuring quality that meets
the needs of all stakeholders. Some participants sustained the argument that higher education
institutions (academics and administrators) should not claim the prerogative to defining quality
outside of the Voice of the Customer. Their role is to work out Voice of the Customer and
through DFSS, SSPD, TFSS, and MFSS design management and products and services design
infrastructure that delivers to the needs and wants of the customers.
The challenge posed by the research findings and validation is the need and feasibility of cross
institutional quality roadmaps. Investing in these roadmaps would benefit even non-QFD quality
initiatives because literature and the Validation Study underscored the need for:
DFSS (Voice of the Customer + Voice of Business)- using DFSS to identify and reduce
what students, employers and venture capitalists feel are deficiencies in the ideal masters
graduate
SSPD (Voice of Employee + Voice of Market) – using SSPD to identify and reduce what
academics, staff and professionals and trendsetters feel are hindrances in quality
performance
TFSS (Voice of Employee + Voice of the Customer) – using TFSS to identify and reduce
what students, academics and staff feel are causes of underperformance in terms of quality
275
MFSS (Voice of Market + Voice of Business) – using MFSS to identify and reduce what
employers, professional bodies, research and the quality assurer feel are roadblocks to
superior quality performance.
Contrary to widely held beliefs about QFD, the model does not, in the service sector, call for
one ‘Quality’. Instead it calls for different ‘quality’ as expressed by Voice of the Customer and
subsequently the forces for Customer Satisfaction Performance. The model fits well a highly
niche-focused university system like the one currently wished-for in Zimbabwe.
6.6 Conclusion
The study has shown that it is feasible to use QFD as a tool for quality assurance of educational
products and services as shown in its management and institutionalisation in the M.Sc.
Programme. The study has indicated a number of implications for knowledge management and
for practice as of policy interest. Recommendations extended to each constituent are carefully
constructed from observations and analysis of fieldwork and literature review. This fits the
discussions within the research problem represented at Figure 6.1from whence I can summarise
the study by saying:
Literature has many fragmented definitions of Quality and small-scale ideas of how the Quality
can be gotten (chapter 2). An inappropriate definition of quality misleads the strategies of
seeking it (chapter 3). These main constrains to quality assurance are exemplified in literature
and discussions (chapter 5) and in chapter 6 it is suggested that only a two-prong approach to
276
QFD as a model for quality assurance can surmount these shortcomings (dual nature of QFD
and the House of Quality).
With respect to Research question 2. There were many motivators to the adoption of QFD.
Multiplicity of stakeholders begets multiplicity of felt-purposes and organisational conflicts. A
deep-going discussion of QFD (chapter 2) shows that these roadblocks to quality performance
are not inherent in Education but are created by those in the practice of educating and managing
the Education systems. The CUTGBS (chapter 5) exemplify this point and show that using
QFD, constraints can be managed away.
With respect to Research question 3. If perceptions and behaviours matter to Quality and quality
assurance in organisations with a multiplicity of constraints (chapter 2) deployment of the 14
management excellence principles (chapter 2) should improve organisational responses to
external publics (chapter 2 / 3) and others’ points of view on Quality.
With respect to Research question 4. QFD tools are the basic units of structure and functioning
of the model (chapter 2 / 3). Laying them out clearly strengthens the model, gives a coherent
shape of what is being built by the model thus begetting approval from the customer.
With respect to Research question 5. Believing that one continues to be part of the problem by
not taking action (chapter 2 + internal quality assurance) and that doing something (strategy
focus wheel) with a model and others (external quality assurance) is a robust approach to quality
assurance.
With respect to Research question 6. Management would respond to any one issue based on
their generic management dispositions. Giving more autonomy to an institution is more about
the impression management gets from those seeking the autonomy as to own management
desires.
.
277
REFERENCES
278
Al-Kassem, A., In’airat, M. and Bakri, A. 2013. Evaluation tools of Total Quality Management
in business organisations. European Journal of Business and Management, 5(6), pp. 41-45.
Almannai, B. Greenough, R. M. and Kay, J. M. 2008. A decision support tool based on QFD
and FMEA for the selection of manufacturing automation technologies. Robotics and
Computer-Integrated Manufacturing, 24(4) pp. 501-507
Alstete, J. W. 1995. Benchmarking in higher education: adapting best practices to improve
quality. Washington D.C.: George Washington University.
Altbach, P. G. 2001. The rise of the pseudo universities. International Higher Education. 25
(Fall). pp. 2-3.
Altbach, P. G. 2009. The Giants Awake: The Present and Future of Higher Education
Systems in China and India. In Higher Education to 2030. Vol. 2 of Globalization, ed.
Organisation for Economic Co-operation and Development
(OECD), 179–204. Paris: OECD.
Altbach, P. G. and Knight, J. 2007. The internationalization of higher education: Motivations
and realities. Journal of Studies in InternationalEducation, 11(3-4), pp. 290-305.
Altbach, G., and Peterson, M. 2007. Higher Education in the New Century: Global
Challenges and Innovative Ideas. Rotterdam, The Netherlands: Sense Publishers.
Altbach, G., Reisberg, L. and Rumbley, L. 2009. Trends in global higher education:
tracking an academic revolution: a report prepared for the UNESCO 2009 World
Conference on Higher Education, UNESCO World Conference on Higher
Education: 2009, 5-8 July, Paris.
Altbach, G. and J. Salmi, J. 2011. The road to academic excellence: The making of world-
class research universities: Washington D.C. The World Bank.
Altbach, G. 2012. Knowledge and education as international commodities: The collapse of the
common good. Accessed at http://www.ejournals.bc.edu/ojs/index.php/ihe/article on 06-02-
2013.
Alvesson, M. and Skoldberg, K. 2000. Reflexive Methodology: New Vistas for Qualitative
Research. 1st ed. Los Angeles. Sage.
Alvesson, M. and Skoldberg, K. 2009. Reflexive Methodology: New Vistas for Qualitative
Research. 2nd ed. Los Angeles. Sage.
279
Anderson, G. 2006. Assuring quality / resisting quality assurance: academics’ responses to
‘quality’ in some Australian universities. Quality in Higher Education, 12(2), pp.161-173.
Arum, R. and Roksa, J. 2011. Academically Adrift: Limited learning on college campuses.
Chicago, IL, Chicago University Press.
Astin, A.W. 1991. Assessment for Excellence. Westport, CT, American Council on
Education / Macmillan.
Aytac, A. and Deniz, V. 2005. Quality Function Deployment in education: A curriculum
review. Quality and Quantity, 39 pp.507-514.
Babakus, E. and Mangold, W. G. 1992. Adapting the SERVQUAL scale to hospital services:
an empirical investigation. Health Services Research, 26(6), pp. 767-86.
Barbour, R. S. 1998. Mixing qualitative methods: Quality assurance or qualitative quagmire?
Qualitative Health Research, 8(3), pp. 352-361.
Bamber, V. and Anderson, S. 2012. Evaluating learning and teaching: institutional
needs and individual practices. International Journal for Academic Development, 17(1),
pp. 5-18.
Bazeley, P. 2013. Qualitative data analysis: Practical strategies. London. Sage.
Becker Assosiates Inc. 2005. Are you hearing voices? Customer loyalty (available at
www.asq.org) accessed 30-10-2010.
Becket, N. and Brooks, M. 2006. Evaluating quality management in university
Departments. Quality Assurance in Education, 14(2), pp. 123-42.
Becket, N. and Brooks, M. 2008. Quality management practice in higher education—what
quality are we actually enhancing? Journal of Hospitality Leisure, Sport & Tourism
Education, 7(1), pp. 40-54.
Bedggood, R. and Pollard, R. 1999. Uses and misuses of student opinion surveys in eight
Australian universities. Australian Journal of Education, 43(2), pp. 129–41.
Bedi, K and Sharma, J. K. 2006. QFD in Business case studies. U21Global working paper
#015, pp. 1-18.
Bennett, R. and Kane, S. 2014. Students’ interpretations of the meanings of questionnaire
items in the National Student Survey. Quality in Higher Education, 20(2), pp. 129-164.
280
Berger, J. B. 2000. Organizational behavior at colleges and student outcomes: a new
Perspective on college impact. Review of Higher Education, 23(2), pp. 177-198.
rgquist, B., Fredriksson, M. and Svensson. M. 2004. TQM-Terrific Quality Marvel or
Tragic Quality Malpractice. In Maria Fredriksson, From customer satisfaction to citizen
interaction: A cooperation model for community development based on TQM. Doctoral
Thesis. Lulea University of Technology. Sweden.
Bernard, H. R. and Ryan, G. W. 2010. Analysing qualitative data: Systematic approaches.
Los Angeles, CA. Sage.
Bernasconi, A. 2011. Private and public pathways to world-class research universities: the
case of Chile. The World Bank.
Berry, T. and Bunning, R. 2006. Leadership. Edinburgh: CAPDM Ltd.
Bertolin, J. and Leite, D. 2008. Quality evaluation of the Brazilian higher education system:
Relevance, diversity, equity and effectiveness. Quality in Higher Education, 14 (2) pp.121-
133.
Bevington, T., and Samson, D. 2012. Implementing Strategic Change: Managing processes
and interfaces to develop a highly productive organisation. London. Sage.
Biglan, A. 1973. The characteristics of subject matter in different academic areas. Journal of
Applied Psychology, 57(3), pp. 195-203.
Bing, S. 2009. Gridlock and solutions: China’s private education in the period of social
transformation. Chinese Education and Society, 42(6), pp. 3-5.
Birnbaum, R. and Deshotels, J. 1999. Has the academy adopted TQM? Planning for Higher
Education, 28(1), pp. 29-37.
Blackmore, J. 2009. Academic pedagogies, quality logics and performative universities:
teaching and what students want. Studies in Higher Education, 34(8), pp. 857-72.
Blackmur, D. 2005. An evaluation of the South African model of MBA accreditation.
Quality in higher education, 11(2) pp. 87-102.
Blackmur, D. 2007. A critical analysis of the UNESCO-OECD guidelines for quality
provision of cross-border higher education. Quality in Higher Education, 13(2), pp. 117-
130.
281
Blackmur, D. 2010. Does the emperor have the right (or any) clothes? The public
regulation of higher education qualities over the last two decades. Quality in
Higher Education, 16(1), pp. 67-69.
Bloor, M., Frankland, J. and Thomas, M. 2001. Focus Group in Social Research. London.
Sage.
Bloom, D., Canning, D. and Chan, K. 2005. Higher Education and Economic
Development in Africa. World Bank. Washington, DC.
Bloxham, S. 2012. You can see the quality in front of your eyes’: grounding academic
standards between rationality and interpretation. Quality in Higher Education, 18 (2) pp.
185-204
Bok, D. 2006. Our Underachieving Colleges, Princeton: Princeton University Press.
Boland, T. and Fowler, T. 2000. A systems perspective of performance management in
public sector organisations. International Journal of Public Sector Management, 13(5), pp.
417-46.
Bolden, R., Hirsh, W., Connor, H., Petrov, G., and Duquemin, A. 2010. Strategies for
effective higher education-employer engagement. (Accessed at http://tinyurl.com/ydnfmu4
on 10 October 2011).
Bolman, L. G. and Deal, T. E. 2008. Reframing Organizations: Artistry, Choice, and
Leadership. Jossey-Bass. San Francisco, CA.
Borden, V. M. H. and Young, J. W. 2008. Measurement validity and accountability for
student learning. New Directions for Institutional Research, S1, pp. 19-37.
Bosch, V. G. and Enriquez, F. T. 2005. TQM and QFD: Exploiting a customer complaint
management system. International Journal of Quality and Reliability Management, 22(1),
pp. 30-37.
Bouchereau, V. and Rowlands, H. 2000. Methods and techniques to help quality function
deployment (QFD), Benchmarking. An International Journal, 7(1) pp. 8-19.
Boud, D. and Falchikov, N. 1989. Quantitative studies of student self-assessment in higher
education: a critical analysis of findings. Higher Education, 18(5), pp. 529-549.
Bowden, J. and Marton, F. 1998. The University of Learning - beyond Quality and
Competence in Higher Education. 1st edition, Kogan Page, London, UK.
282
Bowman, N. A. 2011. Validity of college self-reported gains at diverse institutions.
Educational Researcher, 40(1), pp. 22-24.
Boyce, C. and Neale. 2006. Conducting In-depth Interviews: A guide for designing and
conducting in-depth interviews for evaluation input. Pathfinder International.
Brandenburg U. and Hans de Wit. 2011. The end of internationalisation. International higher
education quarterly, winter 62. www.bc.edu/cihe
Brennan, J. and Shah, T. 2000. Quality assessment and institutional change:
experiences from 14 countries. Higher Education, 40(3), pp. 331-49.
Brickson, S. 2005. Organisational identity orientation: Forging a link between organisational
identity and the organisation’s relations with stakeholders. Administrative Science
Quarterly, 50 pp. 576-809.
Brink, C. 2010. Quality and standards: clarity, comparability and responsibility,
Quality in Higher Education, 16(2), pp. 139-52.
Brinkerhoff, R. O. 2003. The success case method: Find out quickly what’s
working and what’s not. San Francisco. Berrett-Koehler.
Brittingham, B. 2009. Accreditation in the United States: how did we get to where we
are? New Directions for Higher Education, Vol. 145, pp. 7-27.
Broido, E. M. and Manning, K. 2002. Philosophical foundations and current theoretical
perspectives in qualitative research. Journal of College Student Development, 43, pp. 434-
445.
Bryman, A. 1989. Research Methods and Organisation Studies. London. Unwin-Hyman.
Bryson, J. M. and Alston, F. K. 2005. Creating and implementing your strategic plan: A
workbook for public and non-profit organisations. 2nd ed. San Francisco. Jossey-Bass.
Campbell, C. and Rozsnyai, C. 2002. Quality assurance and the development of course
programmes. Bucharest: UNESCO-CEPES.
Caspersen, J., Frølich, N., Karlsen, H. and Aamodt:O. 2014. Learning outcomes across
disciplines and professions: measurement and interpretation. Quality in Higher Education,
20(2), pp. 195-215.
Chan L, and Minj-Lu W. 2002. Quality function deployment: a literature review. European
Journal of Operational Research, 143, pp. 463-497.
283
Chan, C. Y. P. et al. 2006. QFD-based curriculum planning for vocational education. 8th
symposium on QFD. Austin, Texas 2 December 2006.
Chan, Y. P. 2009. QFD-based curriculum planning for vocational education. PhD thesis. The
Hong Kong Polytechnic University, Pao Yue-kong Library.
Chan, L. K. and Wu, M. L. 2003. Quality function deployment: A comprehensive review of
its concept and methods. Quality Engineering, 15(1), pp. 23-35.
Chang, J. F. 2006. Business process management systems: Strategy and implementation.
Auerbach Publications. Boca Raton.
Chang, I. F. and Ku, A. C. H. 1995. Engineering and Technical Education in Taiwan: An
Observation Based on TQM Concept. ASEE, Atlanta, GA.
CHE-HEQC (Council on Higher Education-Higher Education Quality Committee). 2011.
Framework for the second cycle of quality assurance 2012-2017. (Accessed at
http://www.che.ac.za on 02-03-2012).
Chua, C. 2004. Perception of quality in higher education, Proceedings of the Australian
Universities Quality Forum 2004, pp. 1-7, AUQA Occasional Publication.
Clark, T. 2011. Tertiary education in emerging markets: Fertile ground for private investment.
Comparative and Higher Education, 13, Spring, pp. 1-7.
Clayton, M. 1995. QFD – building quality into English universities. 7th symposium on QFD,
Novi, Michigan, Ann Arbor, MI:QFD Institute pp. 171-178.
Coaldrake:1998. The changing climate of Australian higher education: implications for
the UK, paper presented at the Annual Residential Conference of the Committee of Vice-
Chancellors and Principals of the UK, Manchester, September.
Coate, L. E. 1990. TQM at Oregon State University. Journal of Quality and Participation,
(12), pp. 90-101.
Coates, H. 2009. What’s the difference? A model for measuring the value added by higher
education in Australia. Higher Education Management and Policy, 21(1), pp. 2-19.
Coates, H. 2010. Defining and monitoring academic standards in Australian higher
Education. Higher Education Management and Policy, 22(1), pp. 2-17.
Coates, H. and Seifert, T. 2011. Linking assessment for learning, improvement and
accountability. Quality in Higher Education, 17(2), pp. 179-194.
284
Cochrane, T. D. 2014. Critical success factors for transforming pedagogy with mobile Web
2.0. British Journal of Educational Technology, 45(1), pp. 65–82.
Colyvas, J. A. and Powell, W. W. 2006. Roads to institutionalization: the remaking of
boundaries between public and private science, in Staw, B. M. (Ed.) Research in
Organizational Behaviour: An annual series of analytical essays and critical reviews. Oxford.
Elsevier.
Conrad, F. G. and Schober, F. M. 2000. Clarifying question meaning in a household
telephone survey. Public Opinion Quarterly, 64 pp. 1-28.
Considine, M. and Painter, M. (Eds) 1997. Managerialism: the great debate, Carlton.
Melbourne University Press.
Conti, T. 2002. A roadmap through the fog of quality and organisational assessment. Total
Quality Management, 13(8), pp. 1057-1068.
Cooper. 2007. Knowing your ‘lemons’: quality uncertainty in UK higher education.
Quality in Higher Education, 13(1), pp. 19-29.
Crebert, G. 2000. Links between the Purpose and Outcomes of Planning: Perceptions of
heads of school at Griffith University. Journal of Higher Education Policy and
Management, 22(1), pp. 73-84.
Creswell, J. W. and Clark, V. L. 2007. Designing and conducting mixed methods research.
Thousand Oaks, CA: Sage.
Creswell, J. W. and Clark, V. L. 2011. Designing and conducting mixed methods research
2nd ed. Thousand Oaks, CA: Sage.
Crosby, B. 1979. Quality is free: the art of making quality certain. New York, NY: New
American library.
Crotty, M. 1998. The foundations of social research: meaning and perspective in the
research process. Thousand Oaks, CA: Sage.
Crow, K. A. 2014. Customer-focused development with QFD. accessed at
http://www.npd.solutions.com/qfd.html on 06-04-2014.
Cruickshank, M. 2003. Total Quality Management in the higher education sector: A literature
review of the international and Australian perspective. Total Quality Management and
Business Excellence, 14(10), pp. 1159–1167.
285
Cruz-Ruiz, J. S. and Tamayo, F. 2008. QFD application for tackling Internal Customers
Needs as a base for building a Lean Manufacturing System. (n.p.)
Dahlgaard, J. J. and Dahlgaard, S. M. P. 2002. From defect reduction to reduction of waste and
customer / stakeholder satisfaction (understanding the new TQM metrology). Total Quality
Management, 13(8), pp. 1069-1085.
Dale, B. G. 1999. Managing quality. 3rd ed. Oxford. Blackwell Publishers Ltd.
Davenport, T. H. and Short, J. E. 1990. The new industrial engineering: Information
Technology and business process redesign. Sloan Management Review, Summer, pp. 11-27.
Davis, G. F. and Greve, H. R. 1997. Corporate elite networks and governance changes in the
1980s. American Journal of Sociology, 103 pp. 1-37.
Davis, T. L. 2002. Voices of gender role conflict: the social construction of college men’s
identity. Journal of College Student Developmen0074xx, 43, pp. 508-521.
Deal, S. V. and Coverstone-Carroll. 2006. A simplified use of QFD as a system tool for
designing to cost. Available at www.microstat.sm.bmstu.ru-e-library-SysEngineering-
simplif.pdf (accessed on 20-10-2010).
Deem, R. 2001. Globalisation, new managerialism, academic capitalism and
entrepreneurialism in universities: Is the local dimension still important? Comparative
Education, 37(1), pp. 7-20.
Deephouse, D. L. 1996. Does isomorphism legitimate? Academy of Management Journal,
39, pp. 1024-1039.
Deephouse, D. 1999. To be different, or to be the same?: It’s a question (and Theory) of
strategic balance. Strategic Management Journal, 20 pp. 147-166.
Deming, W. E. 1986. Out of the crisis. Cambridge, MA: Massachusetts Institute of
Technology, Center for Advanced Engineering Study.
Denzin, N. K. 1989. The Research Act, volume 3. Prentice Hall. Englewood Cliffs.
Denzin, N. K. and Lincoln, Y. S. 2005. Introduction: The Discipline and Practice of
Qualitative Research. In N. Denzin and Y. Lincoln (eds), Handbook of Qualitative
Research. 2nd ed. Thousand Oaks, CA: Sage.
Denzin, N. K. and Lincoln, Y. S. 2011. The SAGE Handbook of Qualitative Research. 4th ed.
Thousand Oaks, CA: Sage.
286
Dervitsiotis, K. N. 2002. The importance of conversations-for-action for effective strategic
management. Total Quality Management, 13(8), pp. 1087-1098.
Deshpande, R. and Webster, F. E. 1989. Organisational culture and marketing: defining the
research agenda. Journal of Marketing, 53(January) pp.3-15.
Dew, J. 2009. Quality issues in higher education. The Journal of Quality and Participation,
April, pp. 4-9.
Dewi, S. 2011. Privatisation and marketisation of higher education in Indonesia: The
challenge for equal access and academic values. Higher Education, 61 (2), pp. 209-218.
DiCicco-Bloom, B and Crabtree, B. F. 2006. The Qualitative Research Interview. 40, pp.
314-321. Ebsco Publishing online.
Dill, D. D. 2000, Designing academic audit: Lessons learned in Europe and Asia. Quality in
Higher Education, 6(3), pp. 187-207.
Dill, D. D. 2007. Will market competition assure academic quality? An analysis of the UK
and US experience”, in Westerheijden, D.F., Stensaker, B. and Pires da Rosa, M.J. (Eds),
Quality Assurance in Higher Education: Trends in Regulation, Translation and
Transformation, Springer, Berlin.
Dirienzo, C. E., Das, J., Kathryn, T., Cort, K. T. and Burbridge, J. 2007. The effect of
cultural differences in perceptions of transactions costs on national differences in the
preference for international joint ventures. Journal of International Business Studies, 38(2).
Dixon, M. 2006. Globalization and international higher education: Contested positionings.
Journal of Studies in International Education, 10, pp. 319-333.
Dolan, P. 2005. The development of consumer culture, subjectivity and national identity in
Ireland: 1900-1980: PhD Thesis. Goldsmiths College. University of London.
Dolan, P. 2009. Using Documents: A figurational approach, in Hogan, J., Dolan, P. and
Donelly, P. (Eds), Approaches to Quality Research: Theory and its practical application.
Oak Tree Press. Cork.
Dougherty, D., Barnard, H., and Dunne, D. 2005. The rules and resources that generate the
dynamic capability for sustained product innovation. Qualitative Organisational Research,
pp. 37-74.
Douglas, J., Douglas, A. and Barnes, B. 2006. Measuring student satisfaction at a UK
University. Quality Assurance in Education, 14(3), pp. 251-267.
287
Dowling, M. and Henderson, I. S. 2009. Managing personal competences. Edinburgh:
CAPDM Ltd.
Ehlers, U. 2009. Understanding quality culture. Quality Assurance in Education, 17(4), pp. 343-
363.
El-Khawas, E. 2014. Inside quality reform: early results on using outcomes for improvement.
Quality in Higher Education, 20(2), pp. 183-194.
Eley, M. 2001. The course experience questionnaire: altering question format and
phrasing could improve the CEQ’s effectiveness. Higher Education Research
and Development, 20(3), pp. 293-312.
Elias, N. 2000. The Civilising Process: Sociogenetic and Psychogenetic Investigations,
Revised Edition: Oxford. Blackwell Publishers.
Ellis, C. and Berger, L. 2003. ‘Their story / my story / our story: Including the researcher’s
experience in interview research.’ In J.F. Gubrium and J.A. Holstein(eds) Postmodern
Interviewing. Thousand Oaks, CA: Sage.
Emmel, N. 2013. Sampling and choosing cases in qualitative research: a realistic approach.
Los Angeles. Sage.
Eringa, K., and Boer, I. L. J. 1998. Integrating QFD and service blueprinting. Restructuring
educational service processes. Transactions from the 10th symposium on QFD, Novi
Michigan, Ann Arbor. MI.QFD Institute. pp. 289-302.
Ermer, D. S. 1995. Using QFD becomes an educational experience for students and faculty.
Quality Progress, 28(5), pp.131-136.
Etherington, K. 2000. Narrative approaches to working with male survivors of sexual abuse:
The client’s, the counsellor’s and the researcher’s story. London. Jessica Kingsley
Publishers.
Etherington, K. 2004. Becoming a reflexive researcher: Using our selves in research.
London. Jessica Kingsley Publishers.
Ewell. 2007. The ‘Quality Game’: External Review and Institutional Reaction over Three
Decades in the United States”. In: Westerheijden, D.F., Stensaker, B., and Rosa, M.J. (Eds.).
Quality Assurance in Higher Education: Trends in Regulation, Translation and
Transformation. Dordrecht: Springer.
288
Ezell, A. 2007. Accreditation Mills. Washington, DC, American Association of Collegiate
Registrars and Admissions Officers (AACRAO).
Fetterman, D. M. 2001. Foundations of empowerment evaluation. Thousand Oaks. CA: Sage.
Ficalora, J. P. and Cohen, L. 2009. Quality Function Deployment and Six Sigma: A QFD
Handbook. 2nd ed. NJ, Upper Saddle River.
Fielding, N. and Schreier, M. 2001. Introduction: on the comparability between qualitative
and quantitative research methods. Forum: Qualitative Social Research, 2(1) online at
www.qualitative-research.netfgs-texte/1-01/1-01hrsq-e.htm (accessed on 10-12-2012)
File, J., Weko, T., Hauptman, A., Kristensen, B. and Herlitschka, S. 2009. OECD Reviews of
Tertiary Education: Czech Republic, available at www.oecd.org/publishing/corrigenda
accesed on 20-10-2010.
Fill, C. and Jameison, B. 2006. Marketing communications. Edinburgh: CAPDM Ltd.
Fischer, C. T. 2005. Quality Research Methods for Psychologists: introduction through
Empirical Studies. Boston, MA: Academic Press.
Flick, U. 2007. Designing qualitative research. London. Sage
Flick, U.I. and Foster, J. 2007. ‘Social representations’, in C. Willing and W. Stainton-Rogers
(eds), The Sage Handbook of Qualitative Research in Psychology. London: Sage.
Flick, U. 2009. An introduction to qualitative research. 4th ed. London. Sage.
Flumerfelt, S. and Banachowski, M. 2011. Understanding leadership paradigms for
improvement in higher education. Quality Assurance in Education, 19(3), pp. 224-247.
Fontana, A. and Frey, J. H. 2000. The interview: From structured interviews to negotiated
text. In N.K.Denzin and Y.S. Lincoln(eds) Handbook of Qualitative Research (2nd ed.).
Thousand Oaks,CA: Sage.
Franceschini, F. 2002. Advanced Quality Function Deployment. St. Lucie Press. Boca Raton.
Franceschini, F. and Terzago, M. 1998. An application of quality function deployment to
industrial training courses. International Journal of Quality and Reliability Management,
15( 7), pp. 753-768.
Frankfort-Nachmias, C. and Nachmias, D. 1996. Research Methods in the Social Sciences,
5th ed. New York, NY: St. Martin’s Press.
289
Fredriksson, M. 2004. From customer satisfaction to citizen interaction. Doctoral thesis
2004:39, Division of Quality and Environmental Management, Lulea° University of
Technology, Lulea.
Freeman, O. 2009. Analysing Focus Group Data. In J. Hogan, Dolan. and P. Donnelly
(eds.) Approaches to Qualitative Research: Theory and its practical application. A guide for
Dissertation Students. Oak Tree Press.
French, E., Summers, J., Kinash, S., Lawson, R., Taylor, T., Herbert, J., Fallshaw. J. and
Hall, C. 2014. The practice of quality in assuring learning in higher education. Quality in
Higher Education, 20(1), pp. 24-43.
Gallagher, M. 2010. The Accountability for Quality Agenda in Higher Education
(Australia, Group of Eight) Available at: http://www.go8.edu.au/government-abusiness/
go8-policy-a-analysis/2010/238-the-accountability-for-quality-agenda-inhigher-
education (accessed 12-09-2011).
Gallifa, J. and Batalle, P. 2010. Student perceptions of service quality in a multi-campus
higher education system in Spain. Quality Assurance in Education, 18(2), pp. 156-170.
Galvin, T. 2009. ‘‘All for one or one for all’’: Examining collective and comparative
discourse in an organisational name change debate. Qualitative Organisational Research,
pp. 49-89.
Gamage, D.T., Suwanabroma. J., Ueyama, T., Hada, S. and Sekikawa, E. 2008. The impact
of quality assurance measures on student services at the Japanese and Thai private
universities. Quality Assurance in Education, 16(2), pp.181-198.
Garwe, C. E. 2007. Market-orientation of university curricula. Unpublished Master’s degree
dissertation. Chinhoyi University of Technology.
Glen, M. J., Kumar, A. and Cheng, C. H. 1998. QFD application in an educational setting: A
pilot field study. International Journal of Quality and ReliabilityManagement, 13(4), pp. 99-
108.
Glesne, C., and Peshkin, A. 1992. Becoming qualitative researchers. White Plains. NY:
Longman.
Glynn, M. A. 2008. Beyond constraints: How institutions enable identities. In R. Greenwood,
C. Oliver, K. Sahlin, and R. Suddaby (Eds.) The Sage handbook of organisational
institutionalism (pp. 413-430) Thousand Oaks, CA: Sage.
290
Goleman, D. and Senge. 2014. The triple focus: A new approach to education. Florence,
MA: More Than Sound, LLC.
Gonzalez, M., Quesada, G. and Bahill. T. 2003. Improving product design using Quality
Function Deployment: The school furniture case in developing countries. Quality
Engineering Journal, 16(1), pp. 47-58.
Gonzalez, M., Quesada, G., Gourdin, K. and Hartley, M. 2008. Designing a supply chain
management academic curriculum and benchmarking. Quality Assurance in Education,
16(1), pp. 36-60.
Gopinathan, S., and Lee, M. H. 2011. Challenging and co-opting globalisation: Singapore's
strategies in higher education. Journal of Higher Education Policy and Management, 33(3),
pp. 287-299
Graham, P., Lyman, R. and Trow, M. 1995. Accountability of Colleges and Universities:
an essay. New York, The Accountability Study, Colombia University.
Gregorutti, G. 2011. Commercialisation of higher education in Latin America: The case of
Mexico. Comparative and International Higher Education, 3, pp. 11-14.
Griffin, A. and Hauser, J. R. 1993. The Voice of the Customer. Marketing Science, 12(1),
winter, pp. 1-27.
Grimes, M., Malmberg, J., LaBine, G. 1994. Integrating the customers’ voice to improve
educational services through QFD. Transactions from the 7th symposium on QFD, Novi
Michigan, Ann Arbor. MI.QFD Institute. pp. 369-372.
Gubrium, J. F. and Holstein, J. A. 1997. The New Language of Qualitative Method. Oxford:
Oxford University Press.
Gummesson, E. 2000. Qualitative Methods in Management Research. 2nd ed. London. Sage
Publications, Inc.
Gutmann, A. and Thompson, D. 2004. Why deliberative democracy? Princeton University
Press.
Habermas, J. 1993. Justification and application: remarks on discourse ethics. Cambridge –
UK. Polity Press.
Haffez, K. and Mazouz, A. 2011. Using QFD as a higher education management and
governance tool. Current issues of Business and Law, 6(1), pp. 31-52.
291
Haggis, T. 2009. Student learning research: a broader view, in Tight, M., Mok, K.H.,
Huisman, J. and Morphew, C. C. (Eds.) The Routledge International Handbook of Higher
Education . London, Routledge.
Hammersley, M. and Atkinson. 1995. Ethnography: Principles in Practice, 2nd ed.
London. Routledge.
Hartley, J. 2004. Case study research. In C. Cassell and G. Symon, eds. 2004. Essential guide
to qualitative methods in organisational research. London. Sage Publications. pp. 322-333.
Harvey, L. 2002. The end of quality? Quality in Higher Education, 8(1), pp. 5-22.
Harvey, L. 2003. Student feedback. Quality in Higher Education, 9(1), pp. 3-20.
Harvey, L. 2006. Impact of quality assurance: Overview of a discussion between
representatives of external quality assurance agencies. Quality in Higher Education, 12(3)
pp. 287-290.
Harvey, L. and Green, D. 1993. Defining quality. Assessment and Evaluation in Higher
Education. 18(1), pp. 9-34.
Harvey, L. and Knight, P.T. 1996. Transforming Higher Education. Society for Research
Into Higher Education. Open University Press UK.
Harvey, L. and Williams, J. 2010. Fifteen years of quality in higher education. Quality in
Higher Education. 16(1), pp. 3-36.
Harvey, L. and Williams, J. 2010. Fifteen Years of Quality in Higher Education (Part Two).
Quality in Higher Education, 16(2), pp. 81-113
Hay, D. and Fourie, M. 2002. Preparing the way for mergers in South African higher and
further education institutions: an investigation into staff perceptions. Higher Education,
44(1), pp. 115-131.
Hayward, F. M. 2006. Quality assurance and accreditation of higher education in Africa. A
paper prepared for presentation at the conference on higher education reform in
Francophone Africa: Understanding the keys of success. June 13-15. Ouagadougou, Burkina
Faso.
Haworth, J. G. and Conrad, C. F. 1997. Emblems of Quality in Higher Education. Allyn and
Bacon.
Helms, S. and Key, C. H. 1994. Are students more than customers in the classroom? Quality
Progress, 27, pp. 97-99.
292
Henard, F.2010, Learning Our Lesson: Review of quality teaching in higher education.
OECD and Institutional Management in Higher Education. Available
at: http://www.oecd.org/edu/imhe/44058352.pdf (accessed 03-03-2013).
Hertzman, J. and Ackerman, R. 2010. Evaluating quality in associate degree culinary arts
programs, Quality Assurance in Education, 18(3), pp. 205-226.
Heyneman, S. 2011. The corruption of ethics in higher education. international higher
education: The Boston College Centre for International Higher Education, 62 Winter.
Heywood, J. 2000. Assessment in Higher Education: Student learning, teaching,
Programmes and institutions. London. Jessica Kingsley.
Hillman, J. and Plonka, F. 1995. Using QFD for curriculum design. 7th symposium on QFD.
Novi, Michigan, Ann Arbor, MI:QFD Institute, pp.179-182.
Ho, S.K., and Wearn, K. 1995. A TQM model for higher education and training. Training for
Quality, 3(2), pp. 25-33.
Hoegl, M. and Gemuenden, H. G. 2001. Teamwork quality and the success of innovative
projects: a theoretical concept and empirical evidence. Organisational Science, 12(4) July-
August, pp 435-449
Hogan, J., Dolan, P., and Donnelly:(eds.) 2009. Approaches to qualitative research: Theory and
its practical application. A guide for dissertation students. Ireland. Oak Tree Press.
Houston, D. 2007. TQM and higher education: a critical systems perspective on fitness for
Purpose. Quality in Higher Education, 13(1), pp. 3-17.
Houston, D. 2010. Achievements and Consequences of Two Decades of Quality Assurance
in Higher Education: A Personal View From the Edge. Quality in Higher Education, 16(2),
pp. 177-180.
Houston, D. and Maniku, A.A. 2005. Systems perspectives on external quality assurance:
implication for micro-states. Quality in Higher Education, 11(3), pp. 214-226.
Houston, D. and Paewai, S. 2013. Knowledge, power and meanings shaping quality
assurance in higher education: a systemic critique. Quality in Higher Education, 19 (3), pp.
261-282.
Hubbard D.L. 1994. Can higher education learn from factories? Quality Progress, 27(5)
pp. 93-97.
293
Hultgren, F. 1989. Introduction to interpretive inquiry. In E. H. Hultgren and D. L. Coomer
(eds.) Alternative modes of inquiry in home economics research. pp. 37-59. Peoria, IL:
Glencoe.
Hwarng, H. B., and Teo, C. 2001. Translating customers’ voices into operations
requirements: A QFD application in higher education. International Journal of Quality,
18(2), pp. 195-197.
IAR (Instructional Assessment Resources) 2010. Document analysis. Accessed at
http://www.utexas.edu/academic/diia/assessment/iar/teaching/plan/method/doc-analysis.php.
on 13 February 2013.
ICE (International Conference on Education). 2008. National Report on the status of
education in Zimbabwe. 48th session of UNESCO international conference on education.
Geneva, Switzerland 25-28 November 2008.
Ikiz, A. K., Ozdagoglu, G., Erdem, S. and Kaplaseren, F. 2010. Development of educational
structure for business students in quantitative methods area. Accessed at
http://www.qfdi.org/books/symposium on 12-10-2011
IMA (Institute of Management Accountants). 1999. Theory of Constraints (TOC)
Management system fundamentals. Accessed at www.imanet.org on 21-03-2012.
ISO (International Standards Organisation). 2008. Quality management systemsRequirements:
International Organisation for Standards. accessed at https://www.iso.org on 10-11-2012
Ip, B. and Jacobs, G. 2006. QFD for the games industry: results from a practical application.
Total Quality Management, 17(7), pp. 835-856.
Jabnoun, N. 2009. Economic and cultural factors affecting university excellence. Quality
Assurance in Education, 17(4), pp. 416-429.
Jachson, S. 1995. Beyond the fads: systems thinking for managers. Systems Research, 12(1),
pp.25-42.
Jackson, N. 1997. Academic regulation in UK higher education: part II: typologies
and frameworks for discourse and strategic change, Quality Assurance in
Education, 5(3), pp. 165-79.
Jackson, A. Y., and Mazzei, L. A. (eds) 2009. Voice in Qualitative Inquiry: Challenging
conventional, interpretive, and critical concepts in qualitative research. London. Routledge.
294
Jackson, A. Y., and Mazzei, L. A. 2012. Thinking with theory in qualitative research:
Viewing data across multiple perspectives. London. Routledge.
Jaraiedi, M. and Ritz, D. 1994. Total quality management applied to engineering education.
Quality Assurance in Education, 2(1) pp. 32-40
Jauch, L. R. and Orwig, R. A. 1997. A violation of assumptions: Why TQM won’t work in
the Ivory Tower. Journal of Quality Management, 2(2), pp. 279-292.
Jibladze, E. 2013. Running around in circles: quality assurance reforms in Georgia.
Quality in Higher Education, 19(3), pp. 343-357.
Jnanesh, N. A. and Hebbar, K. 2008. Use of QFD analysis in curriculum development of
engineering education and models for curriculum design and delivery. Proceedings of the
World Congress on Engineering and Computer Sciences, San Francisco. October 22-24.
Jones, S. R. and Abes, E. S. 2003. Developing student understanding of HIV-AIDS through
community service-learning: a case study analysis. Journal of College Student
Development, 44, pp. 470-488.
Jones, S. R., Torres, V., and Arminio, J. 2006. Negotiating the complexities of qualitative
research in higher education. Fundamental elements and issues. Routledge, New York.
Jones, S., Lefoe, G., Harvey, M., and Ryland, K. 2012. Distributed leadership: a
collaborative framework for academics, executives and professionals in higher education.
Journal of Higher Education Policy and Management, 34 (1), 67-78.
Jones, M. 2014. Researching organisations: The practice of organisational fieldwork. Los
Angeles, Sage.
Jordens, J.Z. and Zepke, N. 2009. A network approach to curriculum quality assessment.
Quality in Higher Education, 15(3), pp. 279-289.
Juran, J. M. 1993. Made in USA: A renaissance in quality. Harvard business review, pp. 42-50
Kano, N., Seraku, N., Takahashi, F. and Tsuji. S. 1984. Attractive quality and must-be
quality. Japan Society for Control.
Karmel:1990. Higher education—tensions and balance, Journal of Tertiary
Education Administration, 12(2), pp. 329-337.
Karsak, E. E., Sozer, S. and Alptekin, S. E. 2002. Product planning in QFD using a combined
analytic network process and goal programming approach. Computers and Industrial
Engineering, 44, pp. 171-190.
295
Kay, N. 2008. Competitive Strategy. Edinburgh: CAPDM Ltd.
Kee, T. L. 2010. Critical Success Factors (CSFs) for Lean Six Sigma (LSS) Implementation
and its impact on the performance of electronic manufacturing services (EMS) industries.
DBA Thesis
Keller, G. 1993. Strategic planning and management in a competitive environment. New
Directions for Institutional Research, No. 77, pp. 9-16.
Kennedy, M. T. and Fiss, C. 2009. Institutionalization, framing, and diffusion: The logic
of TQM adoption and implementation decisions among U.S. hospitals. Academy of
Management Journal, 52(2), pp. 897-918.
Kezar, A. and Eckel:2002. The effect of institutional culture on change strategies in
higher education. Higher Education, 73(4), pp. 435-460.
Kilbourn, B. 2006. The qualitative doctoral dissertation proposal. Teachers College Record,
108(4), pp. 529-576.
Kimbro, M. B. 2002. A cross-country empirical investigation of corruption and its
relationship to economic, cultural, and monitoring institutions: An examination of the role of
accounting and financial statements. Quality Journal of Accounting, Auditing Finance,
17(4).
Kirkpatrick, D. L. 1994. Evaluating training programs. San Francisco. Berrett-Koehler
Publishers, inc.
Kleijnen, J. Muijtjens, A., Dolmans, D., Willems, J. and Van Hout, H. 2009. Organisational
values in higher education: perceptions and preferences of staff. Quality in Higher
Education, 15(3), pp. 233-249.
Kleijnen, J., Dolmans, D., Willems. J. and Van Hout, H. 2014. Effective quality management
requires a systematic approach and a flexible organisational culture: a qualitative study
among academic staff. Quality in Higher Education, 20(1), pp. 103-126.
Knight, J. 2011. Five Myths about Internationalisation. International Higher Education, 62.
www.bc.edu/cihe
Kohn, A. 1993. Turning learning into a business: concerns about total quality management.
Educational leadership, 51(1), pp. 58-61.
296
Kohoutek, J. (Ed) 2009. Implementation of standards and guidelines for quality assurance in
higher education of Central and Eastern-European countries. Agenda ahead. UNESCO,
Bucharest.
Krishman, M. and Houshman, A. A. 1993. QFD in academia: addressing customer
requirements in the design of engineering curricula. 5th symposium on QFD. Novi,
Michigan, Ann Arbor, MI: QFD Institute, pp. 505-530.
Krucken, G. 2011. A European perspective on new modes of university governance and
actorhood. Research and Occasional Paper Series: CSHE. 17.11.
Kuckartz, U. 2014. Qualitative text analysis: a guide to methods, practice and using
software. London. Sage.
Kvale, S. 1996. InterViews. London. Sage.
Kvale, S. 2007. Doing Interviews. London. Sage.
Lagrosen, S., Seyyed-Hashemi, R., and Leitner, M. 2004. Examination of the dimensions of
quality in higher education. Quality Assurance in Education, 12(2), pp. 61-69.
Lam, K. K. and Zhao, X. 1998. An application of Quality function deployment to improve
the quality of teaching. International Journal of Quality and Reliability
Managememen.15(4-5), pp. 389.
Laurillard, D. 1987. The different forms of learning in psychology and education,
in Richardson, J. T. E., Eysenck, M. W. and Piper, D. W. (Eds.) Student Learning:
Research in education and cognitive psychology. Milton Keynes, Open University Press.
Law, D. C. S. 2010a. Quality assurance in post-secondary education: some common
approaches. Quality Assurance in Education, 18(1), pp. 64-77.
Law, D. 2010b. Quality assurance in post-secondary education: the student experience. Quality
Assurance in Education, 18(4), pp. 250-270.
Law, D. C. S. and Meyer, J. H. F. 2010. Adaptation and validation of the Inventory of
Learning Styles for quality assurance in a Hong Kong post-secondary education
Context. Quality in Higher Education, 16(3), pp. 269-283.
Lemaitre, M. J. 2009. New approaches to quality assurance in the changing world of higher
education. Accessed at http://www.icde.org on 10-07-2012
Lemaitre, M. J. 2012. Quality assurance for higher education: does it make a difference?,
keynote address presented at the INQAAHE Global Forum, Melbourne, Australia, April.
297
Lenn, M. P. 1992. Global trends in quality assurance in higher education. WENR for
professional and international education, 5(2), Spring.
Levin, B. 2013. To know is not enough: Research knowledge and its use, 1(1), pp. 2-31.
Lewis, R. G. and Smith, D. H. 1994. Total Quality in Higher Education, Delray Beach, FL. St
Lucia Press,
Lillis, D. 2012. Systematically evaluating the effectiveness of quality assurance programmes
in leading to improvements in institutional performance. Quality in Higher Education,
18(1), pp. 59-73.
Lin, J., Zhang, Y., Gao, L. and Liu, Y. 2005. Trust, ownership, and autonomy: challenges
facing private higher education in China. The China Review, 5(1), pp. 61-81.
Lincoln, Y. S. and Guba, E. G. 1985. Naturalistic inquiry. Beverly Hills, CA: Sage.
Little, B. and Williams, R. 2010. Students' roles in maintaining quality and in enhancing
learning: Is there a tension?, Quality in Higher Education, 16(2), pp. 115-127.
Litosseliti, L. 2003. Using focus groups in research. London. Continuum.
Lo-Schiavo, F. M., Shatz, M. A. and Poling, D. A. 2008. Strengthening the scholarship
of teaching and learning via experimentation. Teaching of Psychology, 35(4),
pp. 301–04.
Lodge, J. M., and Bonsanquet, A. 2014. Evaluating quality learning in higher
education: re-examining the evidence. Quality in Higher Education, 20(1), pp. 3-23.
Lomas, L. 2002. Does the development of mass education necessarily mean the end of
quality? The 6th QHE seminar. The end of quality? Birmingham, 25-26 May 2001.
Lomas, L. and Nicholls, G. 2005. Enhancing teaching quality through peer review of
teaching. Quality in Higher Education, 11(2) pp. 137-149.
Lomas, L. 2007. Are students customers? Perceptions of academic staff. Quality
in Higher Education, 13(1), pp. 31-44.
Lowden, K., Hall, S., Elliot, D. and Lewin, J. 2011. Employers’ perception of the
employability skills of new graduates. London. Edge Foundation.
McAdam, R., Armstrong, G. and Kelly, B. 1998. Investigation of the relationship between
total quality and innovation: a research study involving small organizations. European
journal of innovation management, 1(3), pp. 139-147.
298
Macfarlane, B. 2012. Intellectual leadership in higher education: renewing the role of the
university professor. London. Routledge.
Macfarlane, B. and Lomas, L. 1999. Stakeholder conceptions of quality in single company
management education. Quality Assurance in Education, 7(2), pp. 77-84.
Macintyre, S. 2004. Universities and the idea of quality, in Carmichael, R. (Ed.)
Proceedings of the Australian Universities Quality Forum 2004: Quality in a
Time of Change, Adelaide, Australia, 7-9 July.
Mahapatra, S. S. and Khan, M. S. 2007. A framework for analyzing quality in education
settings. European Journal of Engineering Education, 32(2), pp. 205-217.
Marginson, S. 2000. Rethinking academic work in the global era. Journal of Higher
Education Policy and Management, 22 (1), pp. 23-35.
Marginson, S. 2009. The knowledge economy and higher education: a system for regulating
the value of knowledge. Higher Education Management and Policy, 21(1), pp. 2-15.
Marshall, C. And Rossman, G. B. 1999. Designing qualitative research. Thousand islands,
CA: Sage Publications.
Marshall, C. and Rossman, G. B. 2006. Designing Qualitative research. 4th ed. Thousand
islands, CA: Sage Publications.
Mason, J. 2002. Qualitative research, 2nd ed. London. Sage.
Massaro, V. 2010. Cui bono? The relevance and impact of quality assurance. Journal of
Higher Education Policy and Management, 32(1), pp. 17-26.
Materu. 2007. Higher education quality assurance in sub-Saharan Africa: status,
challenges, opportunities, and promising practices. World Bank working paper No. 124.
Washington D.C.
Mautsch, L. O., Herzwurm, G., Krams, B. And Schockert, S. 2010. Curriculum planning for
education in enterprise resource planning systems. Accessed at
http://www.qfdi.org/books/symposium on 23-07-2011.
Maykut, P. and Morehouse, R. 2001. Beginning qualitative research. A philosophic and
Practical Guide, London. Falmer Press.
McCabe, J. 2010. Metacognitive awareness of learning strategies in undergraduates. Memory
and Cognition, 39(3), pp. 462-476.
299
McInnis, C. 1998. Change and Continuity in Academic Work. Canberra, Australia,
Department of Employment, Education, Training and Youth Affairs.
McMurty, J. 1991. Education and the Market Model. Journal of Philosophy of Education,
25(2), pp. 209–218.
Meirovich, G. and Romar, E. 2006. The difficulty in implementing TQM in higher education
instruction: The duality of instructor/student roles. Quality Assurance in Education, 14(4),
pp. 324-337.
Meloy, M. 2008. Writing the qualitative research: understanding by doing, 2nd edn.
Mahwah. Lawrence Erlbaum associates, Publishers.
Merriam, S. B. 1998. Qualitative research and case study applications in education. San
Francisco. Jossey-Bass.
Milisiunaite, I. Adomaitiane, R. and Galginaitis, J. 2009. Quality Management as a tool for
quality culture embedment: Vilnius University Approach. Paper presented at the 31st annual
EAIR forum in Vilnius, Lithuania, 23-26 August 2009.
Mizikaci, F. 2006. A systems approach to program evaluation model for quality in higher
education, Quality Assurance in Education, 14(1) pp. 37-53.
Mizuno, S. and Akao, Y. 1994. QFD: The customer-driven approach to quality planning and
deployment, Ed. Asian Productivity Organisation: ISBN 92-833-11221.
Mohamedbhai, G. 2008. The effects of massification on higher education in Africa. Accessed
at www.voced.edu.au/content/ngv63797 on 23-08-2011.
Mok, K, H. and Cheng, A. B. L. 2011. Global aspirations and strategising
for world-class status: new form of politics in higher education governance in Hong Kong.
Journal of Higher Education Policy and Management, 33 (3), pp. 231-251
Morgan, D. L. 1998. The Focus Group Guidebook: Focus Group Kit 1. London. Sage.
Morley, L. 2003. Quality and Power in Higher Education. Philadelphia, PA. Open
University Press.
Morse, J. M. and Richards, L. 2002. ReadMe first for a user’s guide to qualitative methods.
Thousand Oaks. CA: Sage.
Moscovici, S. 1998. The history and actuality of social representations. In U. Flick (ed) The
Psychology of the social. Cambridge. Cambridge University Press, pp. 209-247.
300
Mukaddes, A. M. M., Bagum, M. N., Islam, M. A., Bashar, M. A. and Chakrabarty V. 2010,
Translating the Student’s Voice into Teaching Techniques: A Quality Function Deployment
Approach. Proceedings of the International Conference on Industrial Engineering and
Operations Management, Dhaka, Bangladesh, January 9-10, 2010
Mukherjee, H., and Wong, K. 2011. The National University of Singapore and the
University of Malaya: Common Roots and Different Paths. World Bank Report.
Murgatroyd, S. 1993. The house of quality: Using QFD for instructional design in distance
education. The American Journal of Distance Education, 7, pp. 34-48.
Najafabadi, H. N., Sadeghi, S. and Habibzadeh, P. 2008. TQM in Higher Education Case
Study: Quality in practice at University College of Boras. Master of Science – Engineering
thesis. University College of Boras.
Narang, R. 2012. How do management students perceive the quality of education in public
institutions? Quality assurance in education, 20 (4), pp. 357-371.
Newton, J. 2000. Feeding the beast or improving quality? Academics’ perceptions
of quality assurance and quality monitoring. Quality in Higher Education, 6(2),
pp. 153-163.
Newton, J. 2010. A tale of two ‘qualitys’: Reflections on the quality revolution in higher
Education. Quality in higher education, 16(1), pp. 51-53.
Newton, J. 2012. Is quality assurance leading to enhancement? Accessed at www.gosboo.ru on
12-08-2013.
Nilsson, P., Lofgren, B. Erixon, G. 1995. QFD in the development of Engineering studies.
Transactions from the 7th symposium on QFD, Novi Michigan, Ann Arbor. MI.QFD
Institute. pp. 519-529.
Nunan, T. and Calvert, J. 1992. Report on the project to investigate quality in Distance
Education: National Distance Education Conference.
Nusche, D. 2008. Assessment of Learning Outcomes in Higher Education: A
comparative review of selected practices, OECD Education Working Paper No.
15. Paris, OECD.
Nutt, C. 2002. Why decisions fail. San Francisco. Berrett-Koehler.
Oakland, J. S. 2000. Total Quality Management. Oxford: Butterworth Heinemann.
301
Oakland, J., Tanner, S. and Gadd, K. 2002. Best practice in business excellence. Total
Quality Management, 13(8), pp. 1125-1139.
Okamoto, R. H. and Rioboo, J. C. A. 2002. Deploying and integrating Education system
indicators with QFD – an application case. Transactions from the fourteenth Symposium on
quality function deployment, pp. 93-109.
O’Donoghue, T. 2007. Planning your Qualitative Research Project: An introduction to
interpretivist research in education. London. Routledge.
O’Neill, O. 2002. A question of trust – called to account. Reith Lectures 2002.
Online at: http://www.bbc.co.uk/radio4/reith2002/lecture3.shtml (accessed 02-06-2010).
OECD (Organisation of Economic Cooperation and Development) 2004. Quality and
Recognition in Higher Education: The Cross-border Challenge, OECD Publishing, Paris.
OECD (Organisation of Economic Cooperation and Development) 2008. Tertiary Education for
the Knowledge Society, Vol 1.
Oshiumi, K. 1966. Perfecting Quality Assurance System in Plants. Quality Control, vol. 17
(May), pp. 62-67.
Owlia, M. S. and Aspinwall, E. M. 1997. TQM in higher education –a review. International
Journal of Quality And Reliability Management, 14(5), pp. 527-54.
Owlia, M. S. and Aspinall, E. M. 1998. A framework for measuring quality in engineering
education. Total Quality Management, 9(6), pp. 501-518.
Palmer, S. 2012. Student evaluation of teaching: keeping in touch with reality. Quality in
Higher Education, 18(3), pp. 297-311.
Palmer, D. P., Jennings, D., and Zhou, Z. 1993. Politics and institutional change: Late
adoption of the multidivisional form by large U.S. corporations. Administrative Science
Quarterly, 38, pp. 100-131.
Parker, M. J. and Jary, D. 1995. The McUniversity: organization, management and academic
Subjectivity. Organization: Ethical Values?, 2(2), pp. 310–338.
Paryani, K., Masoudi, A. and Cudney, E. A. 2010. QFD application in the hospitality
industry: A Hotel Case Study. Quality Management Journal, 17(1), pp. 7-28.
Patton, M. Q. 1997. Utilization-focused evaluation: The new century text.Thousand Oaks. CA:
Sage.
302
Patton, M. Q. 2002. Qualitative research and evaluation methods (3rd ed.). Thousand Oaks,
CA: Sage Publications, Inc.
Pearce, J. A. and Robinson, R. B. 2000. Strategic management: formulation, implementation
and control. 7th ed. Boston: McGraw – Hill.
Pearce, J. A. and Robinson, R. B. 2009. Formulation, implementation, and control of
competitive strategies. 11th ed. Boston: McGraw-Hill.
Pierce, J. L. and Newstrom, J. W. 2000. Leaders and the leadership process: readings, self-
assessments and applications. 2nd ed. Boston: McGraw-Hill.
Pompili, G. 2010. Quality in Search of Meanings: The Case of Italian Universities. Quality in
Higher Education, 16(3), pp. 235-245.
Ponterotto, J. G. 2006. Brief Note on the Origins, Evolution, and Meaning of the Qualitative
Research Concept “Thick Description”. The Qualitative Report, 11 (3), pp. 538-549.
Poole, B. 2010. Quality, semantics and the two cultures: Quality Assurance in Education,
18(1), pp. 6-18.
Porter, S. R. 2011. Do college student surveys have any validity? The Review of Higher
Education, 35(1), pp. 45–76.
Postma, D. 2013. Education as change: educational practice and research for transformation.
Education and Change, 18(1) pp. 3-7.
Power, M. 1997. The Audit Society: Rituals of verification. Oxford, Oxford University
Press.
QAA (Quality Assurance Agency). 2009. Thematic enquiries into concerns about academic
quality and standards in Higher Education in England, accessed at
http://www.qaa.ac.uk/standardsandquality/thematicenquiries/FinalReportApril09.pdf on 20-
10-2011.
QAA (Quality Assurance Agency). 2010. Master’s degree characteristics, accessed at
www.qaa.ac.uk on 20-10-2011.
Quinn, A., Lemay, G., Larsen and Johnson, D. M. 2009. Service quality in higher
education. Total Quality Management, 20(2), pp. 139-152
Race. 2006. The Lecturer’s Toolkit: A practical guide to learning, teaching and
assessment. New York, Routledge.
303
Ramírez, G. B. 2013. Studying quality beyond technical rationality: political and symbolic
Perspectives. Quality in Higher Education, 19(2), pp. 126-141
Ramirez, G. B. and Berger, J. B. 2014. Rankings, accreditation, and the international
quest for quality: organising an approach to value in higher education. Quality in Higher
Education, 22(1), pp. 88-104.
Remler, D. K. and Pema, E. 2009. Why Do Institutions of Higher Education Reward
Research While Selling Education? Cambridge, MA, National Bureau of
Economic Research.
Reynolds, A. 1990. ‘Is an external examiner system an adequate guarantee of academic
standards?’ in Loder, C.P.J. (Ed.), Quality Assurance and Accountability in Higher
Education. London, Kogan Page.
Richardson, J. T. E. 2000. Researching Student Learning: Approaches to studying in
campus-based and distance education (Buckingham, SRHE and Open University Press).
Richardson, J. T. E. 2009. What can students’ perceptions of academic quality tell
us?: Research using the Course Experience Questionnaire, in Tight, M., Mok,
K.H., Huisman, J. & Morphew, C.C. (Eds.) The Routledge International
Handbook of Higher Education. London. Routledge.
Roberts, A. and MacLennan, A.F. 2006. Making strategies work. Edinburgh: CAPDM Ltd.
Roberts, A. Wallace, W. and McClure, N. 2005. Strategic risk management. Edinburgh:
CAPDM Ltd.
Roberts, A. Wallace, W. and O’Farrell, P. 2006. Introduction to business research.
Edinburgh: CAPDM Ltd.
Roberts, A. Wallace, W. and Pfab, F. A. 2006. Introduction to business research.
Edinburgh: CAPDM Ltd.
Roberts, A. and Wallace, W. 2009. Project management. Edinburgh: CAPDM Ltd.
Robertson, T. S., Ehashberg, J. and Rymon, T. 1995. New product announcement signals and
incumbent reactions. Journal of Marketing, 59(July), pp. 1-15.
Rossi, H. and Freeman, H. E. and Lipsey, M. W. 1993. Evaluation: A systematic
approach. 6th edn. London: Sage Publications.
Rossouw, D. and van Vuuren, L. 2004. Business Ethics. 3rd ed. Cape Town. Oxford
University Press.
304
Roux, A. L. 2011. The interface between identity and change: How in-service teachers use
discursive strategies to cope with educational change. Education as Change. 15(2), pp 303-
316.
Rowley, J. 1997. Beyond service quality dimensions in higher education and towards a
service contract. Quality Assurance in Education, 5(1), pp. 7-14.
Rowley, D. J., Lujan, H. D. and Dolence, M. G. 1997, Strategic Change in Colleges and
Universities: planning to survive and prosper. San Francisco: Jossey-Bass.
Rowley, D. and Sherman, H. 2002. Implementing the strategic plan. Planning for Higher
Education, 30, pp. 5-14.
Royal, K. D., Gilliland, K. O. and Dent, G. A. 2014. A student-led methodology for
evaluating curricular redundancy. Journal of MultiDisciplinary Evaluation, 10(23) pp. 1-8.
Rubin, H. and Rubin, I. 1995. Qualitative interviewing: The art of hearing data. Thousand
Oaks, CA: Sage.
Rust, R. T., Inman, J., Jia, J., and Zahorik, A. 1999. What you don’t know about customer-
perceived quality: The role of customer expectation distributions. Marketing Science, 18 (1)
pp. 77-92.
Sadlak, J. and Nian Cai, L. (Eds). 2007. The World-Class University and Ranking: Aiming
Beyond Status. Bucharest, UNESCO-CEPES.
Sahlberg, P. 2006. Education reform for raising economic competitiveness. Journal of
Educational Change. Spring.
Sahney, S., Banwet, D., Karunes, S. 2003. Enhancing quality in education: application of
quality function deployment – an industry perspective. Work Study. 52(6), pp. 297-309.
Sahney, S., Banwet, D., and Karunes, S. 2004a. A SERVQUAL and QFD approach to total
quality education: a student perspective. International Journal of Productivity and
Performance Management, 53(2). pp. 143-166.
Sahney, S., Banwet, D. K. and Karunes, S. 2004b. Conceptualising total quality
management in higher education. The TQM Magazine, 16(2), pp. 145-59.
Salih, D., Al-Turki, U. M. and Hawsawi, F. M. 2003. Quality Function Deployment for
designing a basic statistics course. International Journal of Quality & Reliability
Management, 20(6), pp. 740-750.
Sallis, E. 2002. Total Quality Management in education. London. Kogan Page, Ltd.
305
Sallis, E. 2012. Total Quality Management in education. London. Kogan Page, Ltd.
Salmi, J. 2009. The Challenge of Establishing World-Class Universities. Washington, DC:
World Bank.
Salmi, J. 2011a. Nine errors in building a new world-class university. International Higher
Education; The Boston College Centre for International Higher Education, 62(winter), pp.
5-7.
Salmi, J. 2011b. The road to academic excellence: Lessons of experience pp. 323-348, in P. G.
Altbach and J. Salmi The road to academic excellence: The making of world-class research
universities: Washington D.C. The World Bank.
Sanyal, N. R. and Subarna, K. S. 2004. Determinants of bribery in international business.
Thunderbird International Business Review, 46 (2), pp.
Scheinkopf, L. J. 1995. The theory of constraint. Council for continuous improvement. (n.p.).
Sciarelli, S. 2002. Business quality and business ethics. Total Quality Management, 13(8), pp.
1141-1149.
Scott, A. 2008. Strategic Planning. Edinburgh: CAPDM Ltd.
Scott, W. R. 2008. Institutions and Organizations: Ideas and interests, 3rd edn. Thousand
Oaks, Califonia. Sage.
Senge, P., Scharner, C. O., Jaworski, J., and Flowers, B. S. 2005. Presence: Exploring profound
change in people, organisations and society. London. Nicholas Brealey Publishing.
Senge, M., Kleiner, A., Roberts, C., Ross, R. B. and Smith, B. J. 2007. The dance of
change: The challenges of sustaining momentum in learning organisations. London.
Nicholas Brealey Publishing.
Senge, M., Kleiner, A., Roberts, C., Ross, R. B. and Smith, B. J. 2010a. The fifth discipline
fieldbook: Strategies and tools for building a learning organisation. London. Nicholas
Breadley Publishing.
Senge, P., Smith, B., Kruschwitz, N., Laur, J. and Schley, S. 2010b. The necessary revolution:
How individuals and organisations are working together to create a sustainable world.
London. Nicholas Breadley Publishing.
Senge, M., Cambron-McCabe, N., Lucas, T., Smith, B., Dutton, J. and Kleiner, A. 2012.
Schools that learn: A fifth discipline fieldbook for educators, parents, and
everyone who cares about education. London. Nicholas Brealey Publishing.
306
Sevier, R. A. 2003. From strategy to action. University Business, 6(2), pp. 18-19.
Shah, M. 2013. The effectiveness of external quality audits: a study of Australian
Universities. Quality in Higher Education, 19(3), pp. 358-375.
Shah, M. and Nair, C. S. 2014. Turning the ship around: rethinking strategy development and
implementation in universities. Quality Assurance in Education, 22(2), pp. 145-157.
Sharabi, M. 2010. HR manager leadership in quality improvement in a college environment.
Quality Assurance in Education, 18(4), pp. 317-327.
Sharrock, G. 2012. Four management agendas for Australian universities. Journal of Higher
Education Policy and Management, 34(3), pp. 323-337.
Shermis, M. D. 2008. The Collegiate Learning Assessment: a critical perspective. Assessment
Update, 20 (2), pp. 10-12.
Shore, C. and Wright, S. 2000. Coercive accountability: The rise of audit culture in higher
Education, in Strathem, M. (Eds.) Audit Cultures. London. Routledge.
Short, L., Hennessy, M. and Campbell, J. (ch. 5). In Witwer, M. (ed) 1996. Family violence:
Building a coordinated community response: A guide for communities.
Shuib, M. K., Syed, Z., Sharifah, N. A., Abdullah, R., Said, F. and Yusof, R. 2007.
Implementation of quality assurance in public higher education institutions: University of
Malaya experience. Paper presented at Regional Conference on Quality in Higher
Education, Petaling Jaya, Malaysia. December 10-11. Accessed at
http://eprints.um.edu.my/1019 on 23-07-2010.
Silvestro, R. and Westley, C. 2002. Challenging the paradigm of the process enterprise: a
case study analysis of Business Process Reengineering implementation. OMEGA:
International Journal of Management Science, 30, pp 215-225.
Singh, V. Grover, S. and Kumar, A. 2008. Evaluation of quality in an educational institute: a
quality function deployment approach. Educational Research and Review, 3 (4), pp. 162-
168.
Singh, M. 2010. Quality assurance in higher education: which pasts to build on, what
futures to contemplate? Quality in Higher Education, 16(2) pp. 189-194.
Slaughter, S, and Rhoades, G. 2004. Academic capitalism and the new economy: markets,
State, and higher education. Baltimore: Johns Hopkins University Press.
307
Smith, J. A., Flowers, P., and Larkin, M. 2009. Interpretative phenomenological analysis.
London: Sage.
Srinivasan, M., Jones, D. and Miller, A. 2005. Applying theory of constraints principles and
lean thinking at the marine corps maintenance center. Design Acquisition Review Journal.
St. Pierre, E. A. 2009. Afterword: Decentering voice in qualitative inquiry. In Jackson, A. Y.
and Mazzei, L. A. (eds.) Voice in qualitative inquiry. pp. 221-232.
Stake, R. E. 1995. The art of case study research. Thousand Oaks. CA: Sage.
Stake, R. E. 2000. Case studies. In N. K. Denzin and Y. S. Lincoln (eds.) Handbook of
qualitative research (2nd ed. Pp. 435-454) Thousand Oaks. CA: Sage.
Steenkamp, K., Botha, G. and Moloi, K. C. 2012. Sustaining change in a learning organisation.
African Focus Education Review, 9(2), pp 380-394.
Stella, A. and Bhushan, S. 2011. Quality assurance of transnational higher education: the
experiences of Australia and India, accessed at www.auqa.edu.au/files/otherpublications/ on
12-08-2013.
Stensaker, B. 2008. Outcomes of quality assurance: a discussion of knowledge, methodology
and validity. Quality in Higher Education, 14(1), pp. 3-13.
Stensaker, B. and Harvey, L. 2011. Accountability in Higher Education: Global
perspectives on trust and power. New York, Routledge.
Strathern, M. 2000. The tyranny of transparency. British Educational Research
Journal, 26(3), pp. 309-321.
Strauss, A. and Corbin, J. 1990. Basics of qualitative research. Grounded Theory
Procedures and Techniques, Newburry Park, CA: Sage.
Strauss, A. and Corbin, J. 1998. Basics of qualitative research. Thousand Oaks. CA: Sage.
Stufflebeam, D. L. and Shinkfield, A. J. 2007. Evaluation Theory, Models, and Applications,
San Francisco, CA Jossey-Bass.
Szekeres, J. 2004. The invisible workers. Journal of Higher Education Policy and
Management, 26, pp. 7-22.
Tam, M. 2001. Measuring quality and performance in higher education. Quality in Higher
Education, 7(1), pp. 47-54.
Tan, K. C. and Kek, S. W. 2004. Service quality in higher education using an enhanced
SERVQUAL approach. Quality in Higher Education, 10(1), pp. 17-24.
308
Taylor, T., Gough, J., Bundrock, V. and Winter, R. 1998. A bleak outlook: academic
staff perceptions of changes in core activities in Australian higher education. Studies in
Higher Education, 23(3), pp. 255-268.
Tavcar, M. I. 2000. The role of management in fostering quality and excellence in education.
Ways towards quality in education. Proceedings of the international conference. Brdo pri
Kranju, Slovenia, 8-10 April 2000.
Temponi, C. 2005. Continuous improvement framework: implications for academia.
Quality Assurance in Education, Vol. 13 No. 1, pp. 17-36.
Temtime, Z. T. and Mmereki, R. N. 2010. Challenges faced by graduate business education
in Southern Africa: perceptions of MBA participants. Quality assurance in education, 19(2),
pp.110-129.
Tesch, R. 1992. Qualitative research: analysis types and software tools. Hampshire. Falmer
Thakkar, J. Deshmukh, S. G. 2006. Total quality management (TQM) in self-financed
technical institutions A quality function deployment (QFD) and force field analysis
approach. Quality Assurance in Education, 14(1), pp. 54-74.
The World Bank Report. 2000. Higher education in developing countries: Peril and promise.
accessed at http://www.worldbank.org on 10-09-2010
Thijs, A. and Van den Akker, J. 2009. Curriculum development. Enschede. SLO.
Thomas, G. M. 2007. Differentiation, rationalisation and actorhood in world society. Paper
presented at the SGIR conference, Turin, Italy 12-15 september 2007.
Tierney, B. 1998. Responsive University: Restructuring for High Performance. John
Hopkins University Press.
Tonkiss, F. 2004. Analysing Text and Speech: Content and Discourse Analysis. In Seale, C.
(ed.) Researching Society and Culture. Pp. 367-382, London. Sage.
Torres, V. 2004. Familial influences on the identity development of Latino first year
students. Journal of College Student Development, 45, pp. 457-469.
Tran, N. D., Nguyen, T. T. and Nguyen, M. T. N. 2011. The standard of quality for higher
education institutions in Vietnam: a step in the right direction? Quality Assurance in
Education, 19(2), pp. 130-140.
Treacy, M. and Wiersema, F. 1995. The discipline of market leaders. Addison-Wesley.
Reading, MA.
309
Tribus, M. 1990. Three systems of total quality. Tribus. Coral Springs. (n.p.)
Trigwell, K. 2011. ‘Measuring teaching performance’, in Shin, J. C., Toutkoushian, R. K. and
Teichler, U. (Eds.) University Rankings: Theoretical basis, methodology and impacts on
global higher education. London, Springer.
Trivellas and Dargenidou, D. 2009. Leadership and service quality in higher education:
The case of the Technological Educational Institute of Larissa. International Journal of
Quality and Service Sciences, 1(3), pp. 294-310.
Trow, M. 1994. Managerialism and the academic profession: the case of England. Higher
Education Policy, 7(2), pp. 11-18.
Trowler, P. R. 1998. Academics Responding to Change: New higher education frameworks
and academic cultures. Buckingham. Open University Press).
Tsinidou, M., Gerogiannis, V. and Fitsilis. 2010. Evaluation of the factors that determine
quality in higher education: an empirical study. Quality Assurance in Education, 18(3), pp.
227-244.
Udam, M. and Heidmets, M. 2013. Conflicting views on quality: interpretations of ‘a good
university’ by representatives of the state, the market and academia. Quality in Higher
Education, 19(2), 210-224.
UNESCO. 2008. Global Initiative for Quality Assurance Capacity (GIQAC): Governance
Terms, UNESCO, Paris.
Vlasceanu, L. and Barrows, L. C. 2004. Indicators for Institutional and Programme
Accreditation in Higher/Tertiary Education. Bucharest: UNESCO. Available at:
www.cepes.ro/publications/pdf/Indicators.pdf, accessed 06-06-2011.
Vagneur, K. 2008. Corporate governance. Edinburgh: CAPDM Ltd.
Van Manen, M. 1990. Researching lived experience: Human science for an action sensitive
pedagogy. Albany: State University of New York Press.
Van Manen, M. 1990. Researching lived experience: Human science for an action sensitive
pedagogy 2nd ed. London: Althouse Press.
Veiga, A., Rosa, M. J. and Amaral, A. 2011. Understanding the impacts of quality
assessment: an exploratory use of cultural theory. Quality in Higher Education,
17(1), pp. 53-67.
310
Vellnagel, C. C. 2010. Interviewing a way to generate qualitative data: A critical analysis of
pre-crafted interview schedules: essay available at http://www.grin.com
Volkwein, J. F., Lattuca, L. R., Harper, B. J. and Domingo, R. J. 2007. Measuring the impact of
professional accreditation on student experiences and learning outcomes. Research in
Higher Education, 48(2), pp. 251-282.
Vonderembse, M. A. and Raghunathan, T. S. 1997. Quality function deployment’s impact on
product development. International journal of quality sciences, 2(4), pp. 253-271.
Vora, M. K. 2002. Business excellence through quality management. Total Quality
Management, 13(8), pp. 1151-1159.
Vukasovic, M. 2014. Institutionalisation of internal quality assurance: focusing on
institutional work and the significance of disciplinary differences. Quality in Higher
Education, 20(1), pp. 44-63.
Watty, K. 2003. When will academics learn about quality? Quality in Higher Education,
9(3), pp. 213-221.
Wengraf, T. 2001. Qualitative Research Interviewing: Biographic Narrative and Semi-
structured Method. London: Sage.
Wei, W., DeBrot, D. and Witney, C. 2015. The role of leadership in small scale educational
change. Asian Pacific Journal of Education, 35(1), pp40-54.
Whitley, R. 2008. Universities as strategic actors: limitations and variations. Accessed at
www.portlandpress.com/pp/books/online on 27-10-2012.
Wilkes, N. and Dale, B. G. 1998. Attitudes to self-assessment and quality awards: a study in
small and medium-sized companies. Total Quality Management, 9(8), pp. 731-9.
Williams, J. 2011. Editorial. Quality in Higher Education, 17(1), pp. 1-2
Wilson, H. W. 1998. Examining the fabric of academic life: an analysis of three decades of
research on the perception of Australian academics about their roles. Higher Education,
36(4), pp. 421-435.
Wilson, L. 2009. Master’s and doctoral education in Europe: key challenges for Quality
Assurance, Proceedings of the ENQA workshop, Brasov. 12-13 March 2009. Available at:
www.oecd.org/publishing/corrigenda accessed 20-10-2010.
Wu, H. 2009. Prospects of private education in China. Chinese Education and Society,
42(6), pp. 40-60.
311
Yin, R. 1994. Case Study Research: Design and Methods. 2nd ed. Thousand Oaks, Sage.
Yorke, M. 1994. Enhancement-led higher education? Quality Assurance in Education, 2 (3),
pp. 6-12.
Yorke, M. 2000. Developing a quality culture in higher education. Tertiary education and
management, 6 (1), pp. 19-36.
Yorke, M. 2009. Student experience surveys: some methodological considerations
and an empirical evaluation. Assessment and Evaluation in Higher Education, 34(6), pp.
721-739.
Yoshikawa, M., Nitta, K., Watanabe, Y, and Shindo, H. 2014. Improvement of university e-
portfolio system in consideration of students’ demands. Accessed at
http://www.qfdi.org/books/symposium on 12-07-2014.
Zailani, S. and Sasthriyar, S. 2011. Investigation on the Six Sigma critical success factors.
European Journal of Scientific Research, 57(1), pp. 124-132.
Zineldin, M. and Vasicheva, V. 2012. The implementation of TRM philosophy and 5Qs
model in higher education – an exploratory investigation at a Swedish university. Nang Yan
Business Journal, 1(1), pp. 65-75.
Zheng, X. and Pulli, P. 2007. Improving mobile services design: a QFD approach.
Computing and informatics, 26, pp. 369-381
.
312
APPENDICES
An approach
1.a set of detailed methods, procedures and routines created to carry out a specific activity ... www.businessdictionary.com/definition/system
2.a collection of elements or components that are organised for a common purpose www.searchwindowsserver.techtarget.com
3.a group of interacting, interrelated, or interdependent elements forming a complex whole ... www.thefreedictionary.com/system
4.a set of rules, an arrangement of things, or a group of related things that work toward a common goal www.yourdictionary.com/system
1.a study of the general and fundamental problems ... www.wikipedia.org/wiki/philosophy
2.a way of thinking about the world, the universe, and about society www.simple.wikipedia.org/wiki/philosophy
3.a comprehensive system of ideas about human nature and the nature of the reality we live in www.atlassociety.org/what_is_philosophy
4.a group of ideas worked out by a philosopher www.simple.wikipedia.org/wiki/philosophy
A philosophy A system
1.the systematic, theoretical analysis of the methods applied to a field of study www.wikipedia.org/wiki/methodology
2.system of broad principles and or rules from which specific methods or procedures may be derived to interpret or solve different problems
www.businessdictionary.co/.../methodology.html
3.process used to collect information and data for the purpose of making business decisions www.businessdictionary.com/.../research-
methodology
4.is a series of choices www.irn.tufts.edu/research.../methodology_tips.pdf
A methodology
313
1.an established, habitual, logical, or prescribed practice or systematic process of achieving certain ends with accuracy and efficiency
www.businessdictionary.com/.../technique.html
2.a procedure, technique, or way of doing something, especially in accordance with a definite plan www.dictionary.reference.com/browse/method
3.a means or manner of procedure, especially a regular and systematic way of accomplishing something www.thefreedictionary.com/method
4.a particular procedure for accomplishing or approaching something, especially a systematic or established one
www.oxforddictionaries.com/definition/.../method
A method
1.a collection of interrelated work tasks initiated in response to an event that achieves a specific result for the customer of the process
www.irma-international.org/viewtitle/6087/
2.a systematic series of actions directed to some end http://dictionary.reference.com/browse/process
3.a srquence of interdependent and linked procedures which, at every stage, consume one or more resources to convert inputs into outputs
http://.businessdictionary.com/definition/process.html
4.a series of actions, changes, or functions bringing about a result http://www.thefreedictionary.com/process
A process
1.any physical item that can be used to achieve a goal .. www.wikipedia.org/wiki/tool
2.an item or implement used for a specific purpose www.businessdictionary.com/definition/tool.html
3.a method that relate ideas to ideas, ideas to data and data to data (Ficalora & Cohen, 2009:xvii)
4.assist in creative thinking and problem solving (Ficalora & Cohen, 2009:xvii.)
1. a systematic procedure, formula, or routine by which a task is accomplished www.businessdictionary.com/.../technique.html
2. a method of doing some task or performing something www.vocabulary.com/dictionary/technique
3. a way of doing something by using special knowledge or skill Merriam-Webster dictionary programme
4. the method, procedure or way something is done www.yourdictionary.com/technique
A technique A tool
314
Appendix 2: Interview schedule for staff on strategy for quality in the CUTGBS
Date of interview ........................................ Interview with .............................................
1. How would you describe the CUTGBS quality model?
a. TQM /
b. QFD /
c. etc
2. How do you feel staff relate with this model?
a. Enthused
b. Disinterested
c. Indifferent
d. Resisting
e. Learning
3. How has the model faired so far?
a. Disaster
b. Success
c. Struggling
d. Mediocre
4. What was the background to the adoption of the model?
a. Solution looking for a problem
b. Problem looking for a solution
c. Me too strategy
d. Isomorphism
5. What were the assumptions at the adoption of the model and what is the status of these
assumptions now?
6. How successful has been the implementation of the model?
................................................................................................................................................
7. What do you feel stands in the way for successful implementation of the model?
Culture / organisational structure / skills / motivation / leadership / resources.
................................................................................................................................................
................................................................................................................................................
315
Appendix 3: Structure of the M.Sc. Programme
The M.Sc in Strategic Management Programme comprises a taught component followed by a
dissertation. Each candidate is required to take six (6) full courses in Semester 1 as outlined
below:
SEMESTER 1
MSCSM1-01 Quantitative Management.
MSCSM1-02 Strategic Information Management and E- Business.
MSCSM1- 03: Strategic Marketing Management.
MSCSM1-04 Financial Accounting.
MSCSM 1-05 Management of Organizational Behaviour.
MSCSM1-09 Business Research Methods.
MSCSM1-06 Entrepreneurship
SEMESTER 2
Each candidate is required to take four (4) compulsory courses plus two (2) electives in an
area of preferred specialization.
Compulsory courses
MSCSM1-35 Strategic Planning and Management
MSCSM1-11 Strategic Human Resources Management.
MSCSM1-10 Strategic Financial Management.
MSCSM1-33 Economics.
Options/Areas of specialization
A student is expected to elect any two courses in any one area of specialization. Courses on
offer will depend on viability of the numbers and availability of lecturers.
OPTION 1: OPTION 2:
MANAGEMENT STRATEGY MARKETING
316
International Business Management International Purchasing and Supply Chain Management.
MSCSM1-15 MSCSM1-20
OPTION 3: OPTION 4:
FINANCE HUMAN RESOURCES
Tutorials 9 hours
317
Self Directed Learning 27 hours
Assignments 9 hours
Examination 3 hours
SEMESTER 3
Dissertation – MSCSM201
Credits for the Dissertation = 60 from 420 Notional Hours
Breakdown of hours:
Orientation 20 hours
Topic Formulation -Analysis Discussion 80 hours
Proposal writing and supervision 100 hours
Directed supervision. 5 chapters @ 10 hours per chapter 50 hours
Directed library reading 70 hours
Literature review 70 hours
Research Design and Methodology 100 hours
Analysis and Discussion of results 40 hours
Publication of dissertation in part or in full with supervisor 35 hours
Supervisor and student work on dissertation defence 35 hours
Total Notional Hours 600 hours
318
Appendix 4: Interview questions for staff on implementation of QFD stages
1. What aspects of your quality assurance infrastructure or model have been successful?
4. Do you feel the team has successfully protagonised its evolution within the CUTGBS?
5. How has the CUTGBS reacted to the ZimCHE requirements over the years?
2. Who are the people in the team that takes these key management decisions?
3. What other responsibilities are taken by these team members and the ‘team’ as an
institution?
319
4. How would Voice of the Customer be processed?
5. How well did elements of formal design and emergent design of the CUTGBS relate to
1. What has changed over the years in the way the M.Sc. Programme is structured?
2. What has changed over the years in the way courses are designed in the M.Sc.
Programme?
3. How have student responded to the M.Sc. Programme over the years?
4. In what ways have Voice of academics and staff impacted the M.Sc. Programme?
7. In what ways has the industry and society influenced the M.Sc. Programme?
1.1 How have you intervened in shaping the CUTGBS competitive positioning?
320
4. What aspirations have been difficult to realise within the M.Sc. Programme?
5. What tools or instruments do you use to measure the satisfaction of your customers with
4. How supportive is the university’s way of thinking to your reaching these aspirations?
5. How supportive is the university’s processes and procedures to your reaching these
aspirations?
1. How would you explain the fit between university goal infrastructure and the CUTGBS
4. How would you explain the fit between the culture in he CUTGBS and in the mother
university?
321
2. How were the results of such assessments diffused?
1. What things do you think are absolutely important for the mother university?
3. What things do you think are absolutely important for the CUTGBS? What ought to be?
322
Appendix 5: Strategic tools used within QFD (philosophical orientation)
Benchmarking
Delphi Method
Dialectical Inquiry
Environmental Scanning
Experience Curves
Focus Groups
Forecasting
Future Studies
Interviews
Metagame Analysis
Multiple Scenarios
Situational Analysis
Trend Analysis
323
Appendix 6: QFD tools used within QFD’s methodological orientation
Benchmarking
Brainstorming
Capability Analysis
Career-Path Mapping
Check List
Control Chart
Design Of Experiment
Flow Chart
Force-Field Analysis
Histogram
Interrelationship Diagram
Pareto Analysis
Prioritisation Matrix
324
Process Decision Diagram Chart
Process Map
Scatter Plot
Stratification
Tree Diagram
325
Appendix 7: Questions on the link between QFD and external quality mechanisms
1. Do you think that the model you are using (QFD) covers all aspects of quality assurance?
2. How would external quality assurance mechanisms leverage QFD in the CUTGBS?
6. What aspects of external quality management systems do not fit well with the QFD
326
Appendix 8: Interview schedule for staff on management’s response to results of QFD
1. How has management responded to the way you roll-out the quality model in the
CUTGBS?
2. How does the new mode of governance in the CUTGBS impact university management?
3. Do you feel your dependence on the university resembles the traditional dependence of
4. How do you think your model is taken for an examplar by university management?
327
Appendix 9: Interview schedule for the ZimCHE Directors
2. What are the big issues that the ZimCHE faces during the duty of quality assurance?
3. What are your experiences with quality assurance at the Chinhoyi University of
Technology?
4. What have been the trends in the provision of Masters level degrees in our country?
6. What do you think should be the key vectors in their strategic plans?
7. Various institutions have been trying new business based models how is this working?
328
Appendix 10: Interview schedule for students
a. Strategic planning
b. Curriculum design
c. Course design
e. governance
6. What would you consider to be your standards or measures of good quality at Masters
level?
8. How has the M.Sc. Strategic Management Programme faired against your criteria?
9. How would you re-engineer the M.Sc. Strategic Management Programme if given the
powers to do so?
329
Appendix 11: Observation schedule: lecture halls
1. What is the state of the lecture halls / rooms?
......................................................................................................................................................
......................................................................................................................................................
......................................................................................................................................................
2. How are intended activities supported by technology?
......................................................................................................................................................
3. What evidence is there for the use of technology in assisting teaching?
......................................................................................................................................................
......................................................................................................................................................
4. What evidence is there for the use of technology in assisting learning?
......................................................................................................................................................
.....................................................................................................................................................
5. How is technology linked to the ‘Voice of Customer’?
......................................................................................................................................................
......................................................................................................................................................
6. How are seats organised and arranged in the hall / lecturerooms?
......................................................................................................................................................
......................................................................................................................................................
7. What evidence is there of (in)adequacy of seats / space?
......................................................................................................................................................
.............................................................................................................................
8. What evidence is there of atmosphere control?
......................................................................................................................................................
......................................................................................................................................................
9. Are lecturerooms near / far from toilets / water points / smoking areas / other services?
......................................................................................................................................................
......................................................................................................................................................
10. How do structural setups of the lecture halls (en)discourage participation / interaction?
......................................................................................................................................................
330
Appendix 12: Observation schedule: Library
1. Date of observation:......................................................
2. How are books and literature organised?
2.1 By discipline / subject / level
.....................................................................................................................................................
3. What other sources are available?
3.1 Vignettes
................................................................................................................................
3.2 Audios
...................................................................................................................................
3.3 Magazines
...............................................................................................................................
4. What provisions are evident for master’s degree students?
......................................................................................................................................................
5. What search facilities are present in the library?
......................................................................................................................................................
6. What is the state of computers / laptops?
......................................................................................................................................................
7. For what are students using library computers / laptops?
......................................................................................................................................................
8. What evidence is there for (in)adequacy of books / magazines / computers / laptops /
space?
................................................................................................................................................
9. What times does the library open on different days / dates
.............................................................................................................................................
10. What other services are availed through the library?
................................................................................................................................................
331
Appendix 13: Schedule for document analysis
1. Document name:.......................................................................
2. Date of this analysis:................................................................
3. Any previous analysis:
a) ..................................................
b) ...................................................
c) ...................................................
4. Reasons for this analysis:
a) 10 reasons.............................................
b) 20 reasons.............................................
c) Other reasons......................................
5. This document links with:
5.1 Interviews
a) ..................................................
b) ...................................................
5.2 Observations
a) ..................................................
b) .................................................
5.3 Literature
a) .................................................
b) .................................................
6. Summary of points
......................................................................................................................................................
......................................................................................................................................................
......................................................................................................................................................
......................................................................................................................................................
......................................................................................................................................................
......................................................................................................................................................
......................................................................................................................................................
332
Appendix 14: Questions for the Validation Study
1. I have found that the ZimCHE has been facing some issues with quality assurance and
the quality performance of institutions and programmes. Where do these issues start?
2. In my study I found that CUTGBS was instituting a QFD-based model. Do you feel this
3. The QFD model or strategy consists of these 13 stages. Which do you find most
4. My findings are that the CUTGBS was implementing all the QFD stages but with
variable depths and deployment of tools. The greater focus was on Voice of the
5. I found the purported QFD team being made up of academics in the CUTGBS only,
6. I am wondering why some students felt a gap between their expectations and the offered
7. I have seen that the CUTGBS is heavy on Voice of the Customer with the Director
championing this? I don’t see how the CUTGBS is putting the Voice of the Customer
8. I have seen the CUTGBS hassling for greater autonomy. How wrong could I be in this
assertion?
9. I have developed the following model for SWOT analysis. Does it make sense to you?
10. I have summed the emergence of the QFD model in a model like this one. Does this
11. I have built a House of Quality for the CUTGBS with the bricks of your conversations.
333
12. Do you think this model represent your ideas of quality you are chasing in the M.Sc.
Programme?
334
Appendix 15: Letter of consent to participate in the research
Faculty of Education
Department of Science, Mathematics and
Technology Education
Groenkloof Campus
Groenkloof
(012) 420 2207
At .............................................................
335
Appendix 16 Chinhoyi University of Technology Research approval letter
336
Appendix 17 University of Pretoria Ethics clearance certificate
337
Appendix 18 University of Pretoria Integrated declaration form
338